From: John Voltz Date: Fri, 21 Mar 2008 11:21:02 +0000 (-0000) Subject: current avr32 gcc patches break x86 and possibly other systems. moving bad patches... X-Git-Url: https://git.libre-soc.org/?a=commitdiff_plain;h=577cec7bc6d6a1c2b914d0061cdc3cd2ac03e7f3;p=buildroot.git current avr32 gcc patches break x86 and possibly other systems. moving bad patches to separate directory until a proper fix is made. discovered by jacmet and Dr. Nigel Kukard --- diff --git a/toolchain/gcc/4.2.1/900-avr32-atmel.1.3.2.patch b/toolchain/gcc/4.2.1/900-avr32-atmel.1.3.2.patch deleted file mode 100644 index 33ffb9f880..0000000000 --- a/toolchain/gcc/4.2.1/900-avr32-atmel.1.3.2.patch +++ /dev/null @@ -1,22559 +0,0 @@ -diff -Nrup gcc-4.2.1/configure.in gcc-4.2.1.atmel.1.3.2/configure.in ---- gcc-4.2.1/configure.in 2007-05-30 15:48:07.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/configure.in 2007-09-28 10:34:36.000000000 +0200 -@@ -503,6 +503,9 @@ case "${target}" in - arm-*-riscix*) - noconfigdirs="$noconfigdirs ld target-libgloss ${libgcj}" - ;; -+ avr32-*-*) -+ noconfigdirs="$noconfigdirs target-libiberty target-libmudflap target-libffi ${libgcj}" -+ ;; - avr-*-*) - noconfigdirs="$noconfigdirs target-libiberty target-libstdc++-v3 ${libgcj}" - ;; -diff -Nrup gcc-4.2.1/gcc/builtins.c gcc-4.2.1.atmel.1.3.2/gcc/builtins.c ---- gcc-4.2.1/gcc/builtins.c 2007-03-29 18:19:32.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/builtins.c 2007-09-28 10:33:08.000000000 +0200 -@@ -9223,7 +9223,7 @@ validate_arglist (tree arglist, ...) - - do - { -- code = va_arg (ap, enum tree_code); -+ code = va_arg (ap, int); - switch (code) - { - case 0: -diff -Nrup gcc-4.2.1/gcc/calls.c gcc-4.2.1.atmel.1.3.2/gcc/calls.c ---- gcc-4.2.1/gcc/calls.c 2007-06-20 08:44:26.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/calls.c 2007-09-28 10:33:09.000000000 +0200 -@@ -3448,7 +3448,7 @@ emit_library_call_value_1 (int retval, r - for (; count < nargs; count++) - { - rtx val = va_arg (p, rtx); -- enum machine_mode mode = va_arg (p, enum machine_mode); -+ enum machine_mode mode = va_arg (p, int); - - /* We cannot convert the arg value to the mode the library wants here; - must do it earlier where we know the signedness of the arg. */ -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.c gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.c ---- gcc-4.2.1/gcc/config/avr32/avr32.c 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.c 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,7060 @@ -+/* -+ Target hooks and helper functions for AVR32. -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ Initial porting by Anders �dland. -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+#include "config.h" -+#include "system.h" -+#include "coretypes.h" -+#include "tm.h" -+#include "rtl.h" -+#include "tree.h" -+#include "obstack.h" -+#include "regs.h" -+#include "hard-reg-set.h" -+#include "real.h" -+#include "insn-config.h" -+#include "conditions.h" -+#include "output.h" -+#include "insn-attr.h" -+#include "flags.h" -+#include "reload.h" -+#include "function.h" -+#include "expr.h" -+#include "optabs.h" -+#include "toplev.h" -+#include "recog.h" -+#include "ggc.h" -+#include "except.h" -+#include "c-pragma.h" -+#include "integrate.h" -+#include "tm_p.h" -+#include "langhooks.h" -+ -+#include "target.h" -+#include "target-def.h" -+ -+#include -+ -+/* Forward definitions of types. */ -+typedef struct minipool_node Mnode; -+typedef struct minipool_fixup Mfix; -+ -+/* Obstack for minipool constant handling. */ -+static struct obstack minipool_obstack; -+static char *minipool_startobj; -+static rtx minipool_vector_label; -+ -+/* True if we are currently building a constant table. */ -+int making_const_table; -+ -+/* Some forward function declarations */ -+static unsigned long avr32_isr_value (tree); -+static unsigned long avr32_compute_func_type (void); -+static tree avr32_handle_isr_attribute (tree *, tree, tree, int, bool *); -+static tree avr32_handle_acall_attribute (tree *, tree, tree, int, bool *); -+static tree avr32_handle_fndecl_attribute (tree * node, tree name, tree args, -+ int flags, bool * no_add_attrs); -+static void avr32_reorg (void); -+bool avr32_return_in_msb (tree type); -+bool avr32_vector_mode_supported (enum machine_mode mode); -+static void avr32_init_libfuncs (void); -+ -+ -+static void -+avr32_add_gc_roots (void) -+ { -+ gcc_obstack_init (&minipool_obstack); -+ minipool_startobj = (char *) obstack_alloc (&minipool_obstack, 0); -+ } -+ -+ -+/* List of all known AVR32 parts */ -+static const struct part_type_s avr32_part_types[] = { -+ /* name, part_type, architecture type, macro */ -+ {"none", PART_TYPE_AVR32_NONE, ARCH_TYPE_AVR32_AP, "__AVR32__"}, -+ {"ap7000", PART_TYPE_AVR32_AP7000, ARCH_TYPE_AVR32_AP, "__AVR32_AP7000__"}, -+ {"ap7010", PART_TYPE_AVR32_AP7010, ARCH_TYPE_AVR32_AP, "__AVR32_AP7010__"}, -+ {"ap7020", PART_TYPE_AVR32_AP7020, ARCH_TYPE_AVR32_AP, "__AVR32_AP7020__"}, -+ {"uc3a0256", PART_TYPE_AVR32_UC3A0256, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3A0256__"}, -+ {"uc3a0512", PART_TYPE_AVR32_UC3A0512, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3A0512__"}, -+ {"uc3a1128", PART_TYPE_AVR32_UC3A1128, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3A1128__"}, -+ {"uc3a1256", PART_TYPE_AVR32_UC3A1256, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3A1256__"}, -+ {"uc3a1512", PART_TYPE_AVR32_UC3A1512, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3A1512__"}, -+ {"uc3b064", PART_TYPE_AVR32_UC3B064, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B064__"}, -+ {"uc3b0128", PART_TYPE_AVR32_UC3B0128, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B0128__"}, -+ {"uc3b0256", PART_TYPE_AVR32_UC3B0256, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B0256__"}, -+ {"uc3b164", PART_TYPE_AVR32_UC3B164, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B164__"}, -+ {"uc3b1128", PART_TYPE_AVR32_UC3B1128, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B1128__"}, -+ {"uc3b1256", PART_TYPE_AVR32_UC3B1256, ARCH_TYPE_AVR32_UC, -+ "__AVR32_UC3B1256__"}, -+ {NULL, 0, 0, NULL} -+}; -+ -+/* List of all known AVR32 architectures */ -+static const struct arch_type_s avr32_arch_types[] = { -+ /* name, architecture type, microarchitecture type, feature flags, macro */ -+ {"ap", ARCH_TYPE_AVR32_AP, UARCH_TYPE_AVR32B, -+ (FLAG_AVR32_HAS_DSP -+ | FLAG_AVR32_HAS_SIMD -+ | FLAG_AVR32_HAS_UNALIGNED_WORD -+ | FLAG_AVR32_HAS_CACHES -+ | FLAG_AVR32_HAS_BRANCH_PRED -+ | FLAG_AVR32_HAS_RETURN_STACK), -+ "__AVR32_AP__"}, -+ {"uc", ARCH_TYPE_AVR32_UC, UARCH_TYPE_AVR32A, -+ (FLAG_AVR32_HAS_DSP | FLAG_AVR32_HAS_RMW), -+ "__AVR32_UC__"}, -+ {NULL, 0, 0, 0, NULL} -+}; -+ -+/* Default arch name */ -+const char *avr32_arch_name = "ap"; -+const char *avr32_part_name = "none"; -+ -+const struct part_type_s *avr32_part; -+const struct arch_type_s *avr32_arch; -+ -+/* Set default target_flags. */ -+#undef TARGET_DEFAULT_TARGET_FLAGS -+#define TARGET_DEFAULT_TARGET_FLAGS \ -+ (MASK_HAS_ASM_ADDR_PSEUDOS | MASK_MD_REORG_OPTIMIZATION) -+ -+void -+avr32_optimization_options (int level, -+ int size){ -+ if (AVR32_ALWAYS_PIC) -+ flag_pic = 1; -+ -+ /* Enable section anchors if optimization is enabled. */ -+ if (level > 0 || size) -+ flag_section_anchors = 1; -+} -+ -+/* Override command line options */ -+void -+avr32_override_options (void) -+ { -+ const struct part_type_s *part; -+ const struct arch_type_s *arch; -+ -+ /* Check if part type is set. */ -+ for (part = avr32_part_types; part->name; part++) -+ if (strcmp (part->name, avr32_part_name) == 0) -+ break; -+ -+ avr32_part = part; -+ -+ if (!part->name) -+ { -+ fprintf (stderr, "Unknown part `%s' specified\nKnown part names:\n", -+ avr32_part_name); -+ for (part = avr32_part_types; part->name; part++) -+ fprintf (stderr, "\t%s\n", part->name); -+ avr32_part = &avr32_part_types[PART_TYPE_AVR32_NONE]; -+ } -+ -+ avr32_arch = &avr32_arch_types[avr32_part->arch_type]; -+ -+ /* If part was set to "none" then check if arch was set. */ -+ if (strcmp (avr32_part->name, "none") == 0) -+ { -+ /* Check if arch type is set. */ -+ for (arch = avr32_arch_types; arch->name; arch++) -+ if (strcmp (arch->name, avr32_arch_name) == 0) -+ break; -+ -+ avr32_arch = arch; -+ -+ if (!arch->name) -+ { -+ fprintf (stderr, "Unknown arch `%s' specified\nKnown arch names:\n", -+ avr32_arch_name); -+ for (arch = avr32_arch_types; arch->name; arch++) -+ fprintf (stderr, "\t%s\n", arch->name); -+ avr32_arch = &avr32_arch_types[ARCH_TYPE_AVR32_AP]; -+ } -+ } -+ -+ /* If optimization level is two or greater, then align start of loops to a -+ word boundary since this will allow folding the first insn of the loop. -+ Do this only for targets supporting branch prediction. */ -+ if (optimize >= 2 && TARGET_BRANCH_PRED) -+ align_loops = 2; -+ -+ -+ /* Enable section anchors if optimization is enabled. */ -+ if (optimize > 0 || optimize_size) -+ flag_section_anchors = 1; -+ -+ /* Enable fast-float library if unsafe math optimizations -+ are used. */ -+ if (flag_unsafe_math_optimizations) -+ target_flags |= MASK_FAST_FLOAT; -+ -+ /* Check if we should set avr32_imm_in_const_pool -+ based on if caches are present or not. */ -+ if ( avr32_imm_in_const_pool == -1 ) -+ { -+ if ( TARGET_CACHES ) -+ avr32_imm_in_const_pool = 1; -+ else -+ avr32_imm_in_const_pool = 0; -+ } -+ -+ avr32_add_gc_roots (); -+ } -+ -+ -+/* -+If defined, a function that outputs the assembler code for entry to a -+function. The prologue is responsible for setting up the stack frame, -+initializing the frame pointer register, saving registers that must be -+saved, and allocating size additional bytes of storage for the -+local variables. size is an integer. file is a stdio -+stream to which the assembler code should be output. -+ -+The label for the beginning of the function need not be output by this -+macro. That has already been done when the macro is run. -+ -+To determine which registers to save, the macro can refer to the array -+regs_ever_live: element r is nonzero if hard register -+r is used anywhere within the function. This implies the function -+prologue should save register r, provided it is not one of the -+call-used registers. (TARGET_ASM_FUNCTION_EPILOGUE must likewise use -+regs_ever_live.) -+ -+On machines that have ``register windows'', the function entry code does -+not save on the stack the registers that are in the windows, even if -+they are supposed to be preserved by function calls; instead it takes -+appropriate steps to ``push'' the register stack, if any non-call-used -+registers are used in the function. -+ -+On machines where functions may or may not have frame-pointers, the -+function entry code must vary accordingly; it must set up the frame -+pointer if one is wanted, and not otherwise. To determine whether a -+frame pointer is in wanted, the macro can refer to the variable -+frame_pointer_needed. The variable's value will be 1 at run -+time in a function that needs a frame pointer. (see Elimination). -+ -+The function entry code is responsible for allocating any stack space -+required for the function. This stack space consists of the regions -+listed below. In most cases, these regions are allocated in the -+order listed, with the last listed region closest to the top of the -+stack (the lowest address if STACK_GROWS_DOWNWARD is defined, and -+the highest address if it is not defined). You can use a different order -+for a machine if doing so is more convenient or required for -+compatibility reasons. Except in cases where required by standard -+or by a debugger, there is no reason why the stack layout used by GCC -+need agree with that used by other compilers for a machine. -+ */ -+ -+#undef TARGET_ASM_FUNCTION_PROLOGUE -+#define TARGET_ASM_FUNCTION_PROLOGUE avr32_target_asm_function_prologue -+ -+ -+#undef TARGET_DEFAULT_SHORT_ENUMS -+#define TARGET_DEFAULT_SHORT_ENUMS hook_bool_void_false -+ -+#undef TARGET_PROMOTE_FUNCTION_ARGS -+#define TARGET_PROMOTE_FUNCTION_ARGS hook_bool_tree_true -+ -+#undef TARGET_PROMOTE_FUNCTION_RETURN -+#define TARGET_PROMOTE_FUNCTION_RETURN hook_bool_tree_true -+ -+#undef TARGET_PROMOTE_PROTOTYPES -+#define TARGET_PROMOTE_PROTOTYPES hook_bool_tree_true -+ -+#undef TARGET_MUST_PASS_IN_STACK -+#define TARGET_MUST_PASS_IN_STACK avr32_must_pass_in_stack -+ -+#undef TARGET_PASS_BY_REFERENCE -+#define TARGET_PASS_BY_REFERENCE avr32_pass_by_reference -+ -+#undef TARGET_STRICT_ARGUMENT_NAMING -+#define TARGET_STRICT_ARGUMENT_NAMING avr32_strict_argument_naming -+ -+#undef TARGET_VECTOR_MODE_SUPPORTED_P -+#define TARGET_VECTOR_MODE_SUPPORTED_P avr32_vector_mode_supported -+ -+#undef TARGET_RETURN_IN_MEMORY -+#define TARGET_RETURN_IN_MEMORY avr32_return_in_memory -+ -+#undef TARGET_RETURN_IN_MSB -+#define TARGET_RETURN_IN_MSB avr32_return_in_msb -+ -+#undef TARGET_ARG_PARTIAL_BYTES -+#define TARGET_ARG_PARTIAL_BYTES avr32_arg_partial_bytes -+ -+#undef TARGET_STRIP_NAME_ENCODING -+#define TARGET_STRIP_NAME_ENCODING avr32_strip_name_encoding -+ -+#define streq(string1, string2) (strcmp (string1, string2) == 0) -+ -+#undef TARGET_NARROW_VOLATILE_BITFIELD -+#define TARGET_NARROW_VOLATILE_BITFIELD hook_bool_void_false -+ -+#undef TARGET_ATTRIBUTE_TABLE -+#define TARGET_ATTRIBUTE_TABLE avr32_attribute_table -+ -+#undef TARGET_COMP_TYPE_ATTRIBUTES -+#define TARGET_COMP_TYPE_ATTRIBUTES avr32_comp_type_attributes -+ -+ -+#undef TARGET_RTX_COSTS -+#define TARGET_RTX_COSTS avr32_rtx_costs -+ -+#undef TARGET_CANNOT_FORCE_CONST_MEM -+#define TARGET_CANNOT_FORCE_CONST_MEM avr32_cannot_force_const_mem -+ -+#undef TARGET_ASM_INTEGER -+#define TARGET_ASM_INTEGER avr32_assemble_integer -+ -+#undef TARGET_FUNCTION_VALUE -+#define TARGET_FUNCTION_VALUE avr32_function_value -+ -+#undef TARGET_MIN_ANCHOR_OFFSET -+#define TARGET_MIN_ANCHOR_OFFSET (0) -+ -+#undef TARGET_MAX_ANCHOR_OFFSET -+#define TARGET_MAX_ANCHOR_OFFSET ((1 << 15) - 1) -+ -+ -+/* -+ * Switches to the appropriate section for output of constant pool -+ * entry x in mode. You can assume that x is some kind of constant in -+ * RTL. The argument mode is redundant except in the case of a -+ * const_int rtx. Select the section by calling readonly_data_ section -+ * or one of the alternatives for other sections. align is the -+ * constant alignment in bits. -+ * -+ * The default version of this function takes care of putting symbolic -+ * constants in flag_ pic mode in data_section and everything else in -+ * readonly_data_section. -+ */ -+//#undef TARGET_ASM_SELECT_RTX_SECTION -+//#define TARGET_ASM_SELECT_RTX_SECTION avr32_select_rtx_section -+ -+ -+/* -+ * If non-null, this hook performs a target-specific pass over the -+ * instruction stream. The compiler will run it at all optimization -+ * levels, just before the point at which it normally does -+ * delayed-branch scheduling. -+ * -+ * The exact purpose of the hook varies from target to target. Some -+ * use it to do transformations that are necessary for correctness, -+ * such as laying out in-function constant pools or avoiding hardware -+ * hazards. Others use it as an opportunity to do some -+ * machine-dependent optimizations. -+ * -+ * You need not implement the hook if it has nothing to do. The -+ * default definition is null. -+ */ -+#undef TARGET_MACHINE_DEPENDENT_REORG -+#define TARGET_MACHINE_DEPENDENT_REORG avr32_reorg -+ -+/* Target hook for assembling integer objects. -+ Need to handle integer vectors */ -+static bool -+avr32_assemble_integer (rtx x, unsigned int size, int aligned_p) -+ { -+ if (avr32_vector_mode_supported (GET_MODE (x))) -+ { -+ int i, units; -+ -+ if (GET_CODE (x) != CONST_VECTOR) -+ abort (); -+ -+ units = CONST_VECTOR_NUNITS (x); -+ -+ switch (GET_MODE (x)) -+ { -+ case V2HImode: -+ size = 2; -+ break; -+ case V4QImode: -+ size = 1; -+ break; -+ default: -+ abort (); -+ } -+ -+ for (i = 0; i < units; i++) -+ { -+ rtx elt; -+ -+ elt = CONST_VECTOR_ELT (x, i); -+ assemble_integer (elt, size, i == 0 ? 32 : size * BITS_PER_UNIT, 1); -+ } -+ -+ return true; -+ } -+ -+ return default_assemble_integer (x, size, aligned_p); -+ } -+ -+/* -+ * This target hook describes the relative costs of RTL expressions. -+ * -+ * The cost may depend on the precise form of the expression, which is -+ * available for examination in x, and the rtx code of the expression -+ * in which it is contained, found in outer_code. code is the -+ * expression code--redundant, since it can be obtained with GET_CODE -+ * (x). -+ * -+ * In implementing this hook, you can use the construct COSTS_N_INSNS -+ * (n) to specify a cost equal to n fast instructions. -+ * -+ * On entry to the hook, *total contains a default estimate for the -+ * cost of the expression. The hook should modify this value as -+ * necessary. Traditionally, the default costs are COSTS_N_INSNS (5) -+ * for multiplications, COSTS_N_INSNS (7) for division and modulus -+ * operations, and COSTS_N_INSNS (1) for all other operations. -+ * -+ * When optimizing for code size, i.e. when optimize_size is non-zero, -+ * this target hook should be used to estimate the relative size cost -+ * of an expression, again relative to COSTS_N_INSNS. -+ * -+ * The hook returns true when all subexpressions of x have been -+ * processed, and false when rtx_cost should recurse. -+ */ -+ -+/* Worker routine for avr32_rtx_costs. */ -+static inline int -+avr32_rtx_costs_1 (rtx x, enum rtx_code code ATTRIBUTE_UNUSED, -+ enum rtx_code outer ATTRIBUTE_UNUSED) -+ { -+ enum machine_mode mode = GET_MODE (x); -+ -+ switch (GET_CODE (x)) -+ { -+ case MEM: -+ /* Using pre decrement / post increment memory operations on the -+ avr32_uc architecture means that two writebacks must be performed -+ and hence two cycles are needed. */ -+ if (!optimize_size -+ && GET_MODE_SIZE (mode) <= 2 * UNITS_PER_WORD -+ && avr32_arch->arch_type == ARCH_TYPE_AVR32_UC -+ && (GET_CODE (XEXP (x, 0)) == PRE_DEC -+ || GET_CODE (XEXP (x, 0)) == POST_INC)) -+ return COSTS_N_INSNS (5); -+ -+ /* Memory costs quite a lot for the first word, but subsequent words -+ load at the equivalent of a single insn each. */ -+ if (GET_MODE_SIZE (mode) > UNITS_PER_WORD) -+ return COSTS_N_INSNS (3 + (GET_MODE_SIZE (mode) / UNITS_PER_WORD)); -+ -+ return COSTS_N_INSNS (4); -+ case SYMBOL_REF: -+ case CONST: -+ /* These are valid for the pseudo insns: lda.w and call which operates -+ on direct addresses. We assume that the cost of a lda.w is the same -+ as the cost of a ld.w insn. */ -+ return (outer == SET) ? COSTS_N_INSNS (4) : COSTS_N_INSNS (1); -+ case DIV: -+ case MOD: -+ case UDIV: -+ case UMOD: -+ return optimize_size ? COSTS_N_INSNS (1) : COSTS_N_INSNS (16); -+ -+ case ROTATE: -+ case ROTATERT: -+ if (mode == TImode) -+ return COSTS_N_INSNS (100); -+ -+ if (mode == DImode) -+ return COSTS_N_INSNS (10); -+ return COSTS_N_INSNS (4); -+ case ASHIFT: -+ case LSHIFTRT: -+ case ASHIFTRT: -+ case NOT: -+ if (mode == TImode) -+ return COSTS_N_INSNS (10); -+ -+ if (mode == DImode) -+ return COSTS_N_INSNS (4); -+ return COSTS_N_INSNS (1); -+ case PLUS: -+ case MINUS: -+ case NEG: -+ case COMPARE: -+ case ABS: -+ if (GET_MODE_CLASS (mode) == MODE_FLOAT) -+ return COSTS_N_INSNS (100); -+ -+ if (mode == TImode) -+ return COSTS_N_INSNS (50); -+ -+ if (mode == DImode) -+ return COSTS_N_INSNS (2); -+ return COSTS_N_INSNS (1); -+ -+ case MULT: -+ { -+ if (GET_MODE_CLASS (mode) == MODE_FLOAT) -+ return COSTS_N_INSNS (300); -+ -+ if (mode == TImode) -+ return COSTS_N_INSNS (16); -+ -+ if (mode == DImode) -+ return COSTS_N_INSNS (4); -+ -+ if (mode == HImode) -+ return COSTS_N_INSNS (2); -+ -+ return COSTS_N_INSNS (3); -+ } -+ case IF_THEN_ELSE: -+ if (GET_CODE (XEXP (x, 1)) == PC || GET_CODE (XEXP (x, 2)) == PC) -+ return COSTS_N_INSNS (4); -+ return COSTS_N_INSNS (1); -+ case SIGN_EXTEND: -+ case ZERO_EXTEND: -+ /* Sign/Zero extensions of registers cost quite much since these -+ instrcutions only take one register operand which means that gcc -+ often must insert some move instrcutions */ -+ if (mode == QImode || mode == HImode) -+ return (COSTS_N_INSNS (GET_CODE (XEXP (x, 0)) == MEM ? 0 : 1)); -+ return COSTS_N_INSNS (4); -+ case UNSPEC: -+ /* divmod operations */ -+ if (XINT (x, 1) == UNSPEC_UDIVMODSI4_INTERNAL -+ || XINT (x, 1) == UNSPEC_DIVMODSI4_INTERNAL) -+ { -+ return optimize_size ? COSTS_N_INSNS (1) : COSTS_N_INSNS (16); -+ } -+ /* Fallthrough */ -+ default: -+ return COSTS_N_INSNS (1); -+ } -+ } -+ -+static bool -+avr32_rtx_costs (rtx x, int code, int outer_code, int *total) -+ { -+ *total = avr32_rtx_costs_1 (x, code, outer_code); -+ return true; -+ } -+ -+ -+bool -+avr32_cannot_force_const_mem (rtx x ATTRIBUTE_UNUSED) -+ { -+ /* Do not want symbols in the constant pool when compiling pic or if using -+ address pseudo instructions. */ -+ return ((flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS) -+ && avr32_find_symbol (x) != NULL_RTX); -+ } -+ -+ -+/* Table of machine attributes. */ -+const struct attribute_spec avr32_attribute_table[] = { -+ /* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler } */ -+ /* Interrupt Service Routines have special prologue and epilogue -+ requirements. */ -+ {"isr", 0, 1, false, false, false, avr32_handle_isr_attribute}, -+ {"interrupt", 0, 1, false, false, false, avr32_handle_isr_attribute}, -+ {"acall", 0, 1, false, true, true, avr32_handle_acall_attribute}, -+ {"naked", 0, 0, true, false, false, avr32_handle_fndecl_attribute}, -+ {NULL, 0, 0, false, false, false, NULL} -+}; -+ -+ -+typedef struct -+{ -+ const char *const arg; -+ const unsigned long return_value; -+} -+isr_attribute_arg; -+ -+static const isr_attribute_arg isr_attribute_args[] = { -+ {"FULL", AVR32_FT_ISR_FULL}, -+ {"full", AVR32_FT_ISR_FULL}, -+ {"HALF", AVR32_FT_ISR_HALF}, -+ {"half", AVR32_FT_ISR_HALF}, -+ {"NONE", AVR32_FT_ISR_NONE}, -+ {"none", AVR32_FT_ISR_NONE}, -+ {"UNDEF", AVR32_FT_ISR_NONE}, -+ {"undef", AVR32_FT_ISR_NONE}, -+ {"SWI", AVR32_FT_ISR_NONE}, -+ {"swi", AVR32_FT_ISR_NONE}, -+ {NULL, AVR32_FT_ISR_NONE} -+}; -+ -+/* Returns the (interrupt) function type of the current -+ function, or AVR32_FT_UNKNOWN if the type cannot be determined. */ -+ -+static unsigned long -+avr32_isr_value (tree argument) -+ { -+ const isr_attribute_arg *ptr; -+ const char *arg; -+ -+ /* No argument - default to ISR_NONE. */ -+ if (argument == NULL_TREE) -+ return AVR32_FT_ISR_NONE; -+ -+ /* Get the value of the argument. */ -+ if (TREE_VALUE (argument) == NULL_TREE -+ || TREE_CODE (TREE_VALUE (argument)) != STRING_CST) -+ return AVR32_FT_UNKNOWN; -+ -+ arg = TREE_STRING_POINTER (TREE_VALUE (argument)); -+ -+ /* Check it against the list of known arguments. */ -+ for (ptr = isr_attribute_args; ptr->arg != NULL; ptr++) -+ if (streq (arg, ptr->arg)) -+ return ptr->return_value; -+ -+ /* An unrecognized interrupt type. */ -+ return AVR32_FT_UNKNOWN; -+ } -+ -+ -+ -+/* -+These hooks specify assembly directives for creating certain kinds -+of integer object. The TARGET_ASM_BYTE_OP directive creates a -+byte-sized object, the TARGET_ASM_ALIGNED_HI_OP one creates an -+aligned two-byte object, and so on. Any of the hooks may be -+NULL, indicating that no suitable directive is available. -+ -+The compiler will print these strings at the start of a new line, -+followed immediately by the object's initial value. In most cases, -+the string should contain a tab, a pseudo-op, and then another tab. -+ */ -+#undef TARGET_ASM_BYTE_OP -+#define TARGET_ASM_BYTE_OP "\t.byte\t" -+#undef TARGET_ASM_ALIGNED_HI_OP -+#define TARGET_ASM_ALIGNED_HI_OP "\t.align 1\n\t.short\t" -+#undef TARGET_ASM_ALIGNED_SI_OP -+#define TARGET_ASM_ALIGNED_SI_OP "\t.align 2\n\t.int\t" -+#undef TARGET_ASM_ALIGNED_DI_OP -+#define TARGET_ASM_ALIGNED_DI_OP NULL -+#undef TARGET_ASM_ALIGNED_TI_OP -+#define TARGET_ASM_ALIGNED_TI_OP NULL -+#undef TARGET_ASM_UNALIGNED_HI_OP -+#define TARGET_ASM_UNALIGNED_HI_OP "\t.short\t" -+#undef TARGET_ASM_UNALIGNED_SI_OP -+#define TARGET_ASM_UNALIGNED_SI_OP "\t.int\t" -+#undef TARGET_ASM_UNALIGNED_DI_OP -+#define TARGET_ASM_UNALIGNED_DI_OP NULL -+#undef TARGET_ASM_UNALIGNED_TI_OP -+#define TARGET_ASM_UNALIGNED_TI_OP NULL -+ -+#undef TARGET_ASM_OUTPUT_MI_THUNK -+#define TARGET_ASM_OUTPUT_MI_THUNK avr32_output_mi_thunk -+ -+#undef TARGET_ASM_CAN_OUTPUT_MI_THUNK -+#define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_tree_hwi_hwi_tree_true -+ -+static void -+avr32_output_mi_thunk (FILE * file, -+ tree thunk ATTRIBUTE_UNUSED, -+ HOST_WIDE_INT delta, -+ HOST_WIDE_INT vcall_offset, tree function) -+ { -+ int mi_delta = delta; -+ int this_regno = -+ (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function) ? -+ INTERNAL_REGNUM (11) : INTERNAL_REGNUM (12)); -+ -+ -+ if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") -+ || vcall_offset -+ || flag_pic) -+ { -+ fputs ("\tpushm\tlr\n", file); -+ } -+ -+ -+ if (mi_delta != 0) -+ { -+ if (avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21")) -+ { -+ fprintf (file, "\tsub\t%s, -0x%x\n", reg_names[this_regno], -+ mi_delta); -+ } -+ else -+ { -+ /* Immediate is larger than k21 we must make us a temp register by -+ pushing a register to the stack. */ -+ fprintf (file, "\tmov\tlr, lo(%x)\n", mi_delta); -+ fprintf (file, "\torh\tlr, hi(%x)\n", mi_delta); -+ fprintf (file, "\tadd\t%s, lr\n", reg_names[this_regno]); -+ } -+ } -+ -+ -+ if (vcall_offset != 0) -+ { -+ fprintf (file, "\tld.w\tlr, %s[0]\n", reg_names[this_regno]); -+ fprintf (file, "\tld.w\tlr, lr[%i]\n", (int) vcall_offset); -+ fprintf (file, "\tadd\t%s, lr\n", reg_names[this_regno]); -+ } -+ -+ -+ if ( (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") -+ || vcall_offset) -+ && !flag_pic ) -+ { -+ fputs ("\tpopm\tlr\n", file); -+ } -+ -+ if (flag_pic) -+ { -+ /* Load the got into lr and then load the pointer -+ to the function from the got and put it on the stack. -+ We can then call the function and restore lr by issuing -+ a doubleword load from the stack. We do not use a popm/ldm -+ since it will be treated as a return and might need a flushing -+ of the return-stack if available. */ -+ rtx label = gen_label_rtx (); -+ /* Load the got. */ -+ fputs ("\tlddpc\tlr, 0f\n", file); -+ (*targetm.asm_out.internal_label) (file, "L", -+ CODE_LABEL_NUMBER (label)); -+ fputs ("\trsub\tlr, pc\n", file); -+ /* Load the function pointer. */ -+ fputs ("\tld.w\tlr, lr[", file); -+ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); -+ fputs ("@got]\n", file); -+ /* Push the function pointer on the stack.*/ -+ fputs ("\tpushm\tlr\n", file); -+ /* Restore the old lr value and load the function pointer into -+ pc. */ -+ fputs ("\tld.d\tlr,sp++\n", file); -+ fprintf (file, "\t.align 2\n"); -+ fprintf (file, "0:\t.long\t.L%d - _GLOBAL_OFFSET_TABLE_\n", CODE_LABEL_NUMBER (label)); -+ } -+ else -+ { -+ fprintf (file, "\tlddpc\tpc, 0f\n"); -+ fprintf (file, "\t.align 2\n"); -+ fputs ("0:\t.long\t", file); -+ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); -+ fputc ('\n', file); -+ } -+ } -+ -+/* Implements target hook vector_mode_supported. */ -+bool -+avr32_vector_mode_supported (enum machine_mode mode) -+ { -+ if ((mode == V2HImode) || (mode == V4QImode)) -+ return true; -+ -+ return false; -+ } -+ -+ -+#undef TARGET_INIT_LIBFUNCS -+#define TARGET_INIT_LIBFUNCS avr32_init_libfuncs -+ -+#undef TARGET_INIT_BUILTINS -+#define TARGET_INIT_BUILTINS avr32_init_builtins -+ -+#undef TARGET_EXPAND_BUILTIN -+#define TARGET_EXPAND_BUILTIN avr32_expand_builtin -+ -+tree int_ftype_int, int_ftype_void, short_ftype_short, void_ftype_int_int, -+void_ftype_ptr_int; -+tree void_ftype_int, void_ftype_void, int_ftype_ptr_int; -+tree short_ftype_short, int_ftype_int_short, int_ftype_short_short, -+short_ftype_short_short; -+tree int_ftype_int_int, longlong_ftype_int_short, longlong_ftype_short_short; -+tree void_ftype_int_int_int_int_int, void_ftype_int_int_int; -+tree longlong_ftype_int_int, void_ftype_int_int_longlong; -+tree int_ftype_int_int_int, longlong_ftype_longlong_int_short; -+tree longlong_ftype_longlong_short_short, int_ftype_int_short_short; -+ -+#define def_builtin(NAME, TYPE, CODE) \ -+ lang_hooks.builtin_function ((NAME), (TYPE), (CODE), \ -+ BUILT_IN_MD, NULL, NULL_TREE) -+ -+#define def_mbuiltin(MASK, NAME, TYPE, CODE) \ -+ do \ -+ { \ -+ if ((MASK)) \ -+ lang_hooks.builtin_function ((NAME), (TYPE), (CODE), \ -+ BUILT_IN_MD, NULL, NULL_TREE); \ -+ } \ -+ while (0) -+ -+struct builtin_description -+{ -+ const unsigned int mask; -+ const enum insn_code icode; -+ const char *const name; -+ const int code; -+ const enum rtx_code comparison; -+ const unsigned int flag; -+ const tree *ftype; -+}; -+ -+static const struct builtin_description bdesc_2arg[] = { -+#define DSP_BUILTIN(code, builtin, ftype) \ -+ { 1, CODE_FOR_##code, "__builtin_" #code , \ -+ AVR32_BUILTIN_##builtin, 0, 0, ftype } -+ -+ DSP_BUILTIN (mulsathh_h, MULSATHH_H, &short_ftype_short_short), -+ DSP_BUILTIN (mulsathh_w, MULSATHH_W, &int_ftype_short_short), -+ DSP_BUILTIN (mulsatrndhh_h, MULSATRNDHH_H, &short_ftype_short_short), -+ DSP_BUILTIN (mulsatrndwh_w, MULSATRNDWH_W, &int_ftype_int_short), -+ DSP_BUILTIN (mulsatwh_w, MULSATWH_W, &int_ftype_int_short), -+ DSP_BUILTIN (satadd_h, SATADD_H, &short_ftype_short_short), -+ DSP_BUILTIN (satsub_h, SATSUB_H, &short_ftype_short_short), -+ DSP_BUILTIN (satadd_w, SATADD_W, &int_ftype_int_int), -+ DSP_BUILTIN (satsub_w, SATSUB_W, &int_ftype_int_int), -+ DSP_BUILTIN (mulwh_d, MULWH_D, &longlong_ftype_int_short), -+ DSP_BUILTIN (mulnwh_d, MULNWH_D, &longlong_ftype_int_short) -+}; -+ -+ -+void -+avr32_init_builtins (void) -+ { -+ unsigned int i; -+ const struct builtin_description *d; -+ tree endlink = void_list_node; -+ tree int_endlink = tree_cons (NULL_TREE, integer_type_node, endlink); -+ tree longlong_endlink = -+ tree_cons (NULL_TREE, long_long_integer_type_node, endlink); -+ tree short_endlink = -+ tree_cons (NULL_TREE, short_integer_type_node, endlink); -+ tree void_endlink = tree_cons (NULL_TREE, void_type_node, endlink); -+ -+ /* int func (int) */ -+ int_ftype_int = build_function_type (integer_type_node, int_endlink); -+ -+ /* short func (short) */ -+ short_ftype_short -+ = build_function_type (short_integer_type_node, short_endlink); -+ -+ /* short func (short, short) */ -+ short_ftype_short_short -+ = build_function_type (short_integer_type_node, -+ tree_cons (NULL_TREE, short_integer_type_node, -+ short_endlink)); -+ -+ /* long long func (long long, short, short) */ -+ longlong_ftype_longlong_short_short -+ = build_function_type (long_long_integer_type_node, -+ tree_cons (NULL_TREE, long_long_integer_type_node, -+ tree_cons (NULL_TREE, -+ short_integer_type_node, -+ short_endlink))); -+ -+ /* long long func (short, short) */ -+ longlong_ftype_short_short -+ = build_function_type (long_long_integer_type_node, -+ tree_cons (NULL_TREE, short_integer_type_node, -+ short_endlink)); -+ -+ /* int func (int, int) */ -+ int_ftype_int_int -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ int_endlink)); -+ -+ /* long long func (int, int) */ -+ longlong_ftype_int_int -+ = build_function_type (long_long_integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ int_endlink)); -+ -+ /* long long int func (long long, int, short) */ -+ longlong_ftype_longlong_int_short -+ = build_function_type (long_long_integer_type_node, -+ tree_cons (NULL_TREE, long_long_integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ short_endlink))); -+ -+ /* long long int func (int, short) */ -+ longlong_ftype_int_short -+ = build_function_type (long_long_integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ short_endlink)); -+ -+ /* int func (int, short, short) */ -+ int_ftype_int_short_short -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, -+ short_integer_type_node, -+ short_endlink))); -+ -+ /* int func (short, short) */ -+ int_ftype_short_short -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, short_integer_type_node, -+ short_endlink)); -+ -+ /* int func (int, short) */ -+ int_ftype_int_short -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ short_endlink)); -+ -+ /* void func (int, int) */ -+ void_ftype_int_int -+ = build_function_type (void_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ int_endlink)); -+ -+ /* void func (int, int, int) */ -+ void_ftype_int_int_int -+ = build_function_type (void_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ int_endlink))); -+ -+ /* void func (int, int, long long) */ -+ void_ftype_int_int_longlong -+ = build_function_type (void_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ longlong_endlink))); -+ -+ /* void func (int, int, int, int, int) */ -+ void_ftype_int_int_int_int_int -+ = build_function_type (void_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, -+ integer_type_node, -+ tree_cons -+ (NULL_TREE, -+ integer_type_node, -+ int_endlink))))); -+ -+ /* void func (void *, int) */ -+ void_ftype_ptr_int -+ = build_function_type (void_type_node, -+ tree_cons (NULL_TREE, ptr_type_node, int_endlink)); -+ -+ /* void func (int) */ -+ void_ftype_int = build_function_type (void_type_node, int_endlink); -+ -+ /* void func (void) */ -+ void_ftype_void = build_function_type (void_type_node, void_endlink); -+ -+ /* int func (void) */ -+ int_ftype_void = build_function_type (integer_type_node, void_endlink); -+ -+ /* int func (void *, int) */ -+ int_ftype_ptr_int -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, ptr_type_node, int_endlink)); -+ -+ /* int func (int, int, int) */ -+ int_ftype_int_int_int -+ = build_function_type (integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ tree_cons (NULL_TREE, integer_type_node, -+ int_endlink))); -+ -+ /* Initialize avr32 builtins. */ -+ def_builtin ("__builtin_mfsr", int_ftype_int, AVR32_BUILTIN_MFSR); -+ def_builtin ("__builtin_mtsr", void_ftype_int_int, AVR32_BUILTIN_MTSR); -+ def_builtin ("__builtin_mfdr", int_ftype_int, AVR32_BUILTIN_MFDR); -+ def_builtin ("__builtin_mtdr", void_ftype_int_int, AVR32_BUILTIN_MTDR); -+ def_builtin ("__builtin_cache", void_ftype_ptr_int, AVR32_BUILTIN_CACHE); -+ def_builtin ("__builtin_sync", void_ftype_int, AVR32_BUILTIN_SYNC); -+ def_builtin ("__builtin_tlbr", void_ftype_void, AVR32_BUILTIN_TLBR); -+ def_builtin ("__builtin_tlbs", void_ftype_void, AVR32_BUILTIN_TLBS); -+ def_builtin ("__builtin_tlbw", void_ftype_void, AVR32_BUILTIN_TLBW); -+ def_builtin ("__builtin_breakpoint", void_ftype_void, -+ AVR32_BUILTIN_BREAKPOINT); -+ def_builtin ("__builtin_xchg", int_ftype_ptr_int, AVR32_BUILTIN_XCHG); -+ def_builtin ("__builtin_ldxi", int_ftype_ptr_int, AVR32_BUILTIN_LDXI); -+ def_builtin ("__builtin_bswap_16", short_ftype_short, -+ AVR32_BUILTIN_BSWAP16); -+ def_builtin ("__builtin_bswap_32", int_ftype_int, AVR32_BUILTIN_BSWAP32); -+ def_builtin ("__builtin_cop", void_ftype_int_int_int_int_int, -+ AVR32_BUILTIN_COP); -+ def_builtin ("__builtin_mvcr_w", int_ftype_int_int, AVR32_BUILTIN_MVCR_W); -+ def_builtin ("__builtin_mvrc_w", void_ftype_int_int_int, -+ AVR32_BUILTIN_MVRC_W); -+ def_builtin ("__builtin_mvcr_d", longlong_ftype_int_int, -+ AVR32_BUILTIN_MVCR_D); -+ def_builtin ("__builtin_mvrc_d", void_ftype_int_int_longlong, -+ AVR32_BUILTIN_MVRC_D); -+ def_builtin ("__builtin_sats", int_ftype_int_int_int, AVR32_BUILTIN_SATS); -+ def_builtin ("__builtin_satu", int_ftype_int_int_int, AVR32_BUILTIN_SATU); -+ def_builtin ("__builtin_satrnds", int_ftype_int_int_int, -+ AVR32_BUILTIN_SATRNDS); -+ def_builtin ("__builtin_satrndu", int_ftype_int_int_int, -+ AVR32_BUILTIN_SATRNDU); -+ def_builtin ("__builtin_musfr", void_ftype_int, AVR32_BUILTIN_MUSFR); -+ def_builtin ("__builtin_mustr", int_ftype_void, AVR32_BUILTIN_MUSTR); -+ def_builtin ("__builtin_macsathh_w", int_ftype_int_short_short, -+ AVR32_BUILTIN_MACSATHH_W); -+ def_builtin ("__builtin_macwh_d", longlong_ftype_longlong_int_short, -+ AVR32_BUILTIN_MACWH_D); -+ def_builtin ("__builtin_machh_d", longlong_ftype_longlong_short_short, -+ AVR32_BUILTIN_MACHH_D); -+ -+ /* Add all builtins that are more or less simple operations on two -+ operands. */ -+ for (i = 0, d = bdesc_2arg; i < ARRAY_SIZE (bdesc_2arg); i++, d++) -+ { -+ /* Use one of the operands; the target can have a different mode for -+ mask-generating compares. */ -+ -+ if (d->name == 0) -+ continue; -+ -+ def_mbuiltin (d->mask, d->name, *(d->ftype), d->code); -+ } -+ } -+ -+ -+/* Subroutine of avr32_expand_builtin to take care of binop insns. */ -+ -+static rtx -+avr32_expand_binop_builtin (enum insn_code icode, tree arglist, rtx target) -+ { -+ rtx pat; -+ tree arg0 = TREE_VALUE (arglist); -+ tree arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ rtx op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ rtx op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ enum machine_mode tmode = insn_data[icode].operand[0].mode; -+ enum machine_mode mode0 = insn_data[icode].operand[1].mode; -+ enum machine_mode mode1 = insn_data[icode].operand[2].mode; -+ -+ if (!target -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ -+ /* In case the insn wants input operands in modes different from the -+ result, abort. */ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ /* If op0 is already a reg we must cast it to the correct mode. */ -+ if (REG_P (op0)) -+ op0 = convert_to_mode (mode0, op0, 1); -+ else -+ op0 = copy_to_mode_reg (mode0, op0); -+ } -+ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) -+ { -+ /* If op1 is already a reg we must cast it to the correct mode. */ -+ if (REG_P (op1)) -+ op1 = convert_to_mode (mode1, op1, 1); -+ else -+ op1 = copy_to_mode_reg (mode1, op1); -+ } -+ pat = GEN_FCN (icode) (target, op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ } -+ -+/* Expand an expression EXP that calls a built-in function, -+ with result going to TARGET if that's convenient -+ (and in mode MODE if that's convenient). -+ SUBTARGET may be used as the target for computing one of EXP's operands. -+ IGNORE is nonzero if the value is to be ignored. */ -+ -+rtx -+avr32_expand_builtin (tree exp, -+ rtx target, -+ rtx subtarget ATTRIBUTE_UNUSED, -+ enum machine_mode mode ATTRIBUTE_UNUSED, -+ int ignore ATTRIBUTE_UNUSED) -+ { -+ const struct builtin_description *d; -+ unsigned int i; -+ enum insn_code icode; -+ tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0); -+ tree arglist = TREE_OPERAND (exp, 1); -+ tree arg0, arg1, arg2; -+ rtx op0, op1, op2, pat; -+ enum machine_mode tmode, mode0, mode1; -+ enum machine_mode arg0_mode; -+ int fcode = DECL_FUNCTION_CODE (fndecl); -+ -+ switch (fcode) -+ { -+ default: -+ break; -+ -+ case AVR32_BUILTIN_SATS: -+ case AVR32_BUILTIN_SATU: -+ case AVR32_BUILTIN_SATRNDS: -+ case AVR32_BUILTIN_SATRNDU: -+ { -+ const char *fname; -+ switch (fcode) -+ { -+ default: -+ case AVR32_BUILTIN_SATS: -+ icode = CODE_FOR_sats; -+ fname = "sats"; -+ break; -+ case AVR32_BUILTIN_SATU: -+ icode = CODE_FOR_satu; -+ fname = "satu"; -+ break; -+ case AVR32_BUILTIN_SATRNDS: -+ icode = CODE_FOR_satrnds; -+ fname = "satrnds"; -+ break; -+ case AVR32_BUILTIN_SATRNDU: -+ icode = CODE_FOR_satrndu; -+ fname = "satrndu"; -+ break; -+ } -+ -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); -+ -+ tmode = insn_data[icode].operand[0].mode; -+ -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, GET_MODE (op0))) -+ { -+ op0 = copy_to_mode_reg (insn_data[icode].operand[0].mode, op0); -+ } -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) -+ { -+ error ("Parameter 2 to __builtin_%s should be a constant number.", -+ fname); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op2, SImode)) -+ { -+ error ("Parameter 3 to __builtin_%s should be a constant number.", -+ fname); -+ return NULL_RTX; -+ } -+ -+ emit_move_insn (target, op0); -+ pat = GEN_FCN (icode) (target, op1, op2); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return target; -+ } -+ case AVR32_BUILTIN_MUSTR: -+ icode = CODE_FOR_mustr; -+ tmode = insn_data[icode].operand[0].mode; -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ pat = GEN_FCN (icode) (target); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ -+ case AVR32_BUILTIN_MFSR: -+ icode = CODE_FOR_mfsr; -+ arg0 = TREE_VALUE (arglist); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ error ("Parameter 1 to __builtin_mfsr must be a constant number"); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ pat = GEN_FCN (icode) (target, op0); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ case AVR32_BUILTIN_MTSR: -+ icode = CODE_FOR_mtsr; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ mode0 = insn_data[icode].operand[0].mode; -+ mode1 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) -+ { -+ error ("Parameter 1 to __builtin_mtsr must be a constant number"); -+ return gen_reg_rtx (mode0); -+ } -+ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) -+ op1 = copy_to_mode_reg (mode1, op1); -+ pat = GEN_FCN (icode) (op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_MFDR: -+ icode = CODE_FOR_mfdr; -+ arg0 = TREE_VALUE (arglist); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ error ("Parameter 1 to __builtin_mfdr must be a constant number"); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ pat = GEN_FCN (icode) (target, op0); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ case AVR32_BUILTIN_MTDR: -+ icode = CODE_FOR_mtdr; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ mode0 = insn_data[icode].operand[0].mode; -+ mode1 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) -+ { -+ error ("Parameter 1 to __builtin_mtdr must be a constant number"); -+ return gen_reg_rtx (mode0); -+ } -+ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) -+ op1 = copy_to_mode_reg (mode1, op1); -+ pat = GEN_FCN (icode) (op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_CACHE: -+ icode = CODE_FOR_cache; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ mode0 = insn_data[icode].operand[0].mode; -+ mode1 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) -+ { -+ error ("Parameter 2 to __builtin_cache must be a constant number"); -+ return gen_reg_rtx (mode1); -+ } -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) -+ op0 = copy_to_mode_reg (mode0, op0); -+ -+ pat = GEN_FCN (icode) (op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_SYNC: -+ case AVR32_BUILTIN_MUSFR: -+ { -+ const char *fname; -+ switch (fcode) -+ { -+ default: -+ case AVR32_BUILTIN_SYNC: -+ icode = CODE_FOR_sync; -+ fname = "sync"; -+ break; -+ case AVR32_BUILTIN_MUSFR: -+ icode = CODE_FOR_musfr; -+ fname = "musfr"; -+ break; -+ } -+ -+ arg0 = TREE_VALUE (arglist); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ mode0 = insn_data[icode].operand[0].mode; -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) -+ { -+ if (icode == CODE_FOR_musfr) -+ op0 = copy_to_mode_reg (mode0, op0); -+ else -+ { -+ error ("Parameter to __builtin_%s is illegal.", fname); -+ return gen_reg_rtx (mode0); -+ } -+ } -+ pat = GEN_FCN (icode) (op0); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ } -+ case AVR32_BUILTIN_TLBR: -+ icode = CODE_FOR_tlbr; -+ pat = GEN_FCN (icode) (NULL_RTX); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_TLBS: -+ icode = CODE_FOR_tlbs; -+ pat = GEN_FCN (icode) (NULL_RTX); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_TLBW: -+ icode = CODE_FOR_tlbw; -+ pat = GEN_FCN (icode) (NULL_RTX); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_BREAKPOINT: -+ icode = CODE_FOR_breakpoint; -+ pat = GEN_FCN (icode) (NULL_RTX); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return NULL_RTX; -+ case AVR32_BUILTIN_XCHG: -+ icode = CODE_FOR_sync_lock_test_and_setsi; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ mode1 = insn_data[icode].operand[2].mode; -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) -+ { -+ op1 = copy_to_mode_reg (mode1, op1); -+ } -+ -+ op0 = gen_rtx_MEM (SImode, op0); -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ error -+ ("Parameter 1 to __builtin_xchg must be a pointer to an integer."); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ pat = GEN_FCN (icode) (target, op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ case AVR32_BUILTIN_LDXI: -+ icode = CODE_FOR_ldxi; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ mode1 = insn_data[icode].operand[2].mode; -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ op0 = copy_to_mode_reg (mode0, op0); -+ } -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) -+ { -+ op1 = copy_to_mode_reg (mode1, op1); -+ } -+ -+ if (!(*insn_data[icode].operand[3].predicate) (op2, SImode)) -+ { -+ error -+ ("Parameter 3 to __builtin_ldxi must be a valid extract shift operand: (0|8|16|24)"); -+ return gen_reg_rtx (mode0); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ pat = GEN_FCN (icode) (target, op0, op1, op2); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ case AVR32_BUILTIN_BSWAP16: -+ { -+ icode = CODE_FOR_bswap_16; -+ arg0 = TREE_VALUE (arglist); -+ arg0_mode = TYPE_MODE (TREE_TYPE (arg0)); -+ mode0 = insn_data[icode].operand[1].mode; -+ if (arg0_mode != mode0) -+ arg0 = build1 (NOP_EXPR, -+ (*lang_hooks.types.type_for_mode) (mode0, 0), arg0); -+ -+ op0 = expand_expr (arg0, NULL_RTX, HImode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ op0 = copy_to_mode_reg (mode0, op0); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ { -+ target = gen_reg_rtx (tmode); -+ } -+ -+ -+ pat = GEN_FCN (icode) (target, op0); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return target; -+ } -+ case AVR32_BUILTIN_BSWAP32: -+ { -+ icode = CODE_FOR_bswap_32; -+ arg0 = TREE_VALUE (arglist); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) -+ { -+ op0 = copy_to_mode_reg (mode0, op0); -+ } -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ -+ -+ pat = GEN_FCN (icode) (target, op0); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return target; -+ } -+ case AVR32_BUILTIN_MVCR_W: -+ case AVR32_BUILTIN_MVCR_D: -+ { -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ -+ if (fcode == AVR32_BUILTIN_MVCR_W) -+ icode = CODE_FOR_mvcrsi; -+ else -+ icode = CODE_FOR_mvcrdi; -+ -+ tmode = insn_data[icode].operand[0].mode; -+ -+ if (target == 0 -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op0, SImode)) -+ { -+ error -+ ("Parameter 1 to __builtin_cop is not a valid coprocessor number."); -+ error ("Number should be between 0 and 7."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op1, SImode)) -+ { -+ error -+ ("Parameter 2 to __builtin_cop is not a valid coprocessor register number."); -+ error ("Number should be between 0 and 15."); -+ return NULL_RTX; -+ } -+ -+ pat = GEN_FCN (icode) (target, op0, op1); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return target; -+ } -+ case AVR32_BUILTIN_MACSATHH_W: -+ case AVR32_BUILTIN_MACWH_D: -+ case AVR32_BUILTIN_MACHH_D: -+ { -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); -+ -+ icode = ((fcode == AVR32_BUILTIN_MACSATHH_W) ? CODE_FOR_macsathh_w : -+ (fcode == AVR32_BUILTIN_MACWH_D) ? CODE_FOR_macwh_d : -+ CODE_FOR_machh_d); -+ -+ tmode = insn_data[icode].operand[0].mode; -+ mode0 = insn_data[icode].operand[1].mode; -+ mode1 = insn_data[icode].operand[2].mode; -+ -+ -+ if (!target -+ || GET_MODE (target) != tmode -+ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) -+ target = gen_reg_rtx (tmode); -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, tmode)) -+ { -+ /* If op0 is already a reg we must cast it to the correct mode. */ -+ if (REG_P (op0)) -+ op0 = convert_to_mode (tmode, op0, 1); -+ else -+ op0 = copy_to_mode_reg (tmode, op0); -+ } -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op1, mode0)) -+ { -+ /* If op1 is already a reg we must cast it to the correct mode. */ -+ if (REG_P (op1)) -+ op1 = convert_to_mode (mode0, op1, 1); -+ else -+ op1 = copy_to_mode_reg (mode0, op1); -+ } -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op2, mode1)) -+ { -+ /* If op1 is already a reg we must cast it to the correct mode. */ -+ if (REG_P (op2)) -+ op2 = convert_to_mode (mode1, op2, 1); -+ else -+ op2 = copy_to_mode_reg (mode1, op2); -+ } -+ -+ emit_move_insn (target, op0); -+ -+ pat = GEN_FCN (icode) (target, op1, op2); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ return target; -+ } -+ case AVR32_BUILTIN_MVRC_W: -+ case AVR32_BUILTIN_MVRC_D: -+ { -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); -+ -+ if (fcode == AVR32_BUILTIN_MVRC_W) -+ icode = CODE_FOR_mvrcsi; -+ else -+ icode = CODE_FOR_mvrcdi; -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, SImode)) -+ { -+ error ("Parameter 1 is not a valid coprocessor number."); -+ error ("Number should be between 0 and 7."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) -+ { -+ error ("Parameter 2 is not a valid coprocessor register number."); -+ error ("Number should be between 0 and 15."); -+ return NULL_RTX; -+ } -+ -+ if (GET_CODE (op2) == CONST_INT -+ || GET_CODE (op2) == CONST -+ || GET_CODE (op2) == SYMBOL_REF || GET_CODE (op2) == LABEL_REF) -+ { -+ op2 = force_const_mem (insn_data[icode].operand[2].mode, op2); -+ } -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op2, GET_MODE (op2))) -+ op2 = copy_to_mode_reg (insn_data[icode].operand[2].mode, op2); -+ -+ -+ pat = GEN_FCN (icode) (op0, op1, op2); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return NULL_RTX; -+ } -+ case AVR32_BUILTIN_COP: -+ { -+ rtx op3, op4; -+ tree arg3, arg4; -+ icode = CODE_FOR_cop; -+ arg0 = TREE_VALUE (arglist); -+ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); -+ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); -+ arg3 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (TREE_CHAIN (arglist)))); -+ arg4 = -+ TREE_VALUE (TREE_CHAIN -+ (TREE_CHAIN (TREE_CHAIN (TREE_CHAIN (arglist))))); -+ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); -+ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); -+ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); -+ op3 = expand_expr (arg3, NULL_RTX, VOIDmode, 0); -+ op4 = expand_expr (arg4, NULL_RTX, VOIDmode, 0); -+ -+ if (!(*insn_data[icode].operand[0].predicate) (op0, SImode)) -+ { -+ error -+ ("Parameter 1 to __builtin_cop is not a valid coprocessor number."); -+ error ("Number should be between 0 and 7."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) -+ { -+ error -+ ("Parameter 2 to __builtin_cop is not a valid coprocessor register number."); -+ error ("Number should be between 0 and 15."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[2].predicate) (op2, SImode)) -+ { -+ error -+ ("Parameter 3 to __builtin_cop is not a valid coprocessor register number."); -+ error ("Number should be between 0 and 15."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[3].predicate) (op3, SImode)) -+ { -+ error -+ ("Parameter 4 to __builtin_cop is not a valid coprocessor register number."); -+ error ("Number should be between 0 and 15."); -+ return NULL_RTX; -+ } -+ -+ if (!(*insn_data[icode].operand[4].predicate) (op4, SImode)) -+ { -+ error -+ ("Parameter 5 to __builtin_cop is not a valid coprocessor operation."); -+ error ("Number should be between 0 and 127."); -+ return NULL_RTX; -+ } -+ -+ pat = GEN_FCN (icode) (op0, op1, op2, op3, op4); -+ if (!pat) -+ return 0; -+ emit_insn (pat); -+ -+ return target; -+ } -+ } -+ -+ for (i = 0, d = bdesc_2arg; i < ARRAY_SIZE (bdesc_2arg); i++, d++) -+ if (d->code == fcode) -+ return avr32_expand_binop_builtin (d->icode, arglist, target); -+ -+ -+ /* @@@ Should really do something sensible here. */ -+ return NULL_RTX; -+ } -+ -+ -+/* Handle an "interrupt" or "isr" attribute; -+ arguments as in struct attribute_spec.handler. */ -+ -+static tree -+avr32_handle_isr_attribute (tree * node, tree name, tree args, -+ int flags, bool * no_add_attrs) -+ { -+ if (DECL_P (*node)) -+ { -+ if (TREE_CODE (*node) != FUNCTION_DECL) -+ { -+ warning ("`%s' attribute only applies to functions", -+ IDENTIFIER_POINTER (name)); -+ *no_add_attrs = true; -+ } -+ /* FIXME: the argument if any is checked for type attributes; should it -+ be checked for decl ones? */ -+ } -+ else -+ { -+ if (TREE_CODE (*node) == FUNCTION_TYPE -+ || TREE_CODE (*node) == METHOD_TYPE) -+ { -+ if (avr32_isr_value (args) == AVR32_FT_UNKNOWN) -+ { -+ warning ("`%s' attribute ignored", IDENTIFIER_POINTER (name)); -+ *no_add_attrs = true; -+ } -+ } -+ else if (TREE_CODE (*node) == POINTER_TYPE -+ && (TREE_CODE (TREE_TYPE (*node)) == FUNCTION_TYPE -+ || TREE_CODE (TREE_TYPE (*node)) == METHOD_TYPE) -+ && avr32_isr_value (args) != AVR32_FT_UNKNOWN) -+ { -+ *node = build_variant_type_copy (*node); -+ TREE_TYPE (*node) = build_type_attribute_variant -+ (TREE_TYPE (*node), -+ tree_cons (name, args, TYPE_ATTRIBUTES (TREE_TYPE (*node)))); -+ *no_add_attrs = true; -+ } -+ else -+ { -+ /* Possibly pass this attribute on from the type to a decl. */ -+ if (flags & ((int) ATTR_FLAG_DECL_NEXT -+ | (int) ATTR_FLAG_FUNCTION_NEXT -+ | (int) ATTR_FLAG_ARRAY_NEXT)) -+ { -+ *no_add_attrs = true; -+ return tree_cons (name, args, NULL_TREE); -+ } -+ else -+ { -+ warning ("`%s' attribute ignored", IDENTIFIER_POINTER (name)); -+ } -+ } -+ } -+ -+ return NULL_TREE; -+ } -+ -+/* Handle an attribute requiring a FUNCTION_DECL; -+ arguments as in struct attribute_spec.handler. */ -+static tree -+avr32_handle_fndecl_attribute (tree * node, tree name, -+ tree args ATTRIBUTE_UNUSED, -+ int flags ATTRIBUTE_UNUSED, -+ bool * no_add_attrs) -+ { -+ if (TREE_CODE (*node) != FUNCTION_DECL) -+ { -+ warning ("%qs attribute only applies to functions", -+ IDENTIFIER_POINTER (name)); -+ *no_add_attrs = true; -+ } -+ -+ return NULL_TREE; -+ } -+ -+ -+/* Handle an acall attribute; -+ arguments as in struct attribute_spec.handler. */ -+ -+static tree -+avr32_handle_acall_attribute (tree * node, tree name, -+ tree args ATTRIBUTE_UNUSED, -+ int flags ATTRIBUTE_UNUSED, bool * no_add_attrs) -+ { -+ if (TREE_CODE (*node) == FUNCTION_TYPE || TREE_CODE (*node) == METHOD_TYPE) -+ { -+ warning ("`%s' attribute not yet supported...", -+ IDENTIFIER_POINTER (name)); -+ *no_add_attrs = true; -+ return NULL_TREE; -+ } -+ -+ warning ("`%s' attribute only applies to functions", -+ IDENTIFIER_POINTER (name)); -+ *no_add_attrs = true; -+ return NULL_TREE; -+ } -+ -+ -+/* Return 0 if the attributes for two types are incompatible, 1 if they -+ are compatible, and 2 if they are nearly compatible (which causes a -+ warning to be generated). */ -+ -+static int -+avr32_comp_type_attributes (tree type1, tree type2) -+ { -+ int acall1, acall2, isr1, isr2, naked1, naked2; -+ -+ /* Check for mismatch of non-default calling convention. */ -+ if (TREE_CODE (type1) != FUNCTION_TYPE) -+ return 1; -+ -+ /* Check for mismatched call attributes. */ -+ acall1 = lookup_attribute ("acall", TYPE_ATTRIBUTES (type1)) != NULL; -+ acall2 = lookup_attribute ("acall", TYPE_ATTRIBUTES (type2)) != NULL; -+ naked1 = lookup_attribute ("naked", TYPE_ATTRIBUTES (type1)) != NULL; -+ naked2 = lookup_attribute ("naked", TYPE_ATTRIBUTES (type2)) != NULL; -+ isr1 = lookup_attribute ("isr", TYPE_ATTRIBUTES (type1)) != NULL; -+ if (!isr1) -+ isr1 = lookup_attribute ("interrupt", TYPE_ATTRIBUTES (type1)) != NULL; -+ -+ isr2 = lookup_attribute ("isr", TYPE_ATTRIBUTES (type2)) != NULL; -+ if (!isr2) -+ isr2 = lookup_attribute ("interrupt", TYPE_ATTRIBUTES (type2)) != NULL; -+ -+ if ((acall1 && isr2) -+ || (acall2 && isr1) || (naked1 && isr2) || (naked2 && isr1)) -+ return 0; -+ -+ return 1; -+ } -+ -+ -+/* Computes the type of the current function. */ -+ -+static unsigned long -+avr32_compute_func_type (void) -+ { -+ unsigned long type = AVR32_FT_UNKNOWN; -+ tree a; -+ tree attr; -+ -+ if (TREE_CODE (current_function_decl) != FUNCTION_DECL) -+ abort (); -+ -+ /* Decide if the current function is volatile. Such functions never -+ return, and many memory cycles can be saved by not storing register -+ values that will never be needed again. This optimization was added to -+ speed up context switching in a kernel application. */ -+ if (optimize > 0 -+ && TREE_NOTHROW (current_function_decl) -+ && TREE_THIS_VOLATILE (current_function_decl)) -+ type |= AVR32_FT_VOLATILE; -+ -+ if (cfun->static_chain_decl != NULL) -+ type |= AVR32_FT_NESTED; -+ -+ attr = DECL_ATTRIBUTES (current_function_decl); -+ -+ a = lookup_attribute ("isr", attr); -+ if (a == NULL_TREE) -+ a = lookup_attribute ("interrupt", attr); -+ -+ if (a == NULL_TREE) -+ type |= AVR32_FT_NORMAL; -+ else -+ type |= avr32_isr_value (TREE_VALUE (a)); -+ -+ -+ a = lookup_attribute ("acall", attr); -+ if (a != NULL_TREE) -+ type |= AVR32_FT_ACALL; -+ -+ a = lookup_attribute ("naked", attr); -+ if (a != NULL_TREE) -+ type |= AVR32_FT_NAKED; -+ -+ return type; -+ } -+ -+/* Returns the type of the current function. */ -+ -+static unsigned long -+avr32_current_func_type (void) -+ { -+ if (AVR32_FUNC_TYPE (cfun->machine->func_type) == AVR32_FT_UNKNOWN) -+ cfun->machine->func_type = avr32_compute_func_type (); -+ -+ return cfun->machine->func_type; -+ } -+ -+/* -+ This target hook should return true if we should not pass type solely -+ in registers. The file expr.h defines a definition that is usually appropriate, -+ refer to expr.h for additional documentation. -+ */ -+bool -+avr32_must_pass_in_stack (enum machine_mode mode ATTRIBUTE_UNUSED, tree type) -+ { -+ if (type && AGGREGATE_TYPE_P (type) -+ /* If the alignment is less than the size then pass in the struct on -+ the stack. */ -+ && ((unsigned int) TYPE_ALIGN_UNIT (type) < -+ (unsigned int) int_size_in_bytes (type)) -+ /* If we support unaligned word accesses then structs of size 4 and 8 -+ can have any alignment and still be passed in registers. */ -+ && !(TARGET_UNALIGNED_WORD -+ && (int_size_in_bytes (type) == 4 -+ || int_size_in_bytes (type) == 8)) -+ /* Double word structs need only a word alignment. */ -+ && !(int_size_in_bytes (type) == 8 && TYPE_ALIGN_UNIT (type) >= 4)) -+ return true; -+ -+ if (type && AGGREGATE_TYPE_P (type) -+ /* Structs of size 3,5,6,7 are always passed in registers. */ -+ && (int_size_in_bytes (type) == 3 -+ || int_size_in_bytes (type) == 5 -+ || int_size_in_bytes (type) == 6 || int_size_in_bytes (type) == 7)) -+ return true; -+ -+ -+ return (type && TREE_ADDRESSABLE (type)); -+ } -+ -+ -+bool -+avr32_strict_argument_naming (CUMULATIVE_ARGS * ca ATTRIBUTE_UNUSED) -+ { -+ return true; -+ } -+ -+/* -+ This target hook should return true if an argument at the position indicated -+ by cum should be passed by reference. This predicate is queried after target -+ independent reasons for being passed by reference, such as TREE_ADDRESSABLE (type). -+ -+ If the hook returns true, a copy of that argument is made in memory and a -+ pointer to the argument is passed instead of the argument itself. The pointer -+ is passed in whatever way is appropriate for passing a pointer to that type. -+ */ -+bool -+avr32_pass_by_reference (CUMULATIVE_ARGS * cum ATTRIBUTE_UNUSED, -+ enum machine_mode mode ATTRIBUTE_UNUSED, -+ tree type, bool named ATTRIBUTE_UNUSED) -+ { -+ return (type && (TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)); -+ } -+ -+static int -+avr32_arg_partial_bytes (CUMULATIVE_ARGS * pcum ATTRIBUTE_UNUSED, -+ enum machine_mode mode ATTRIBUTE_UNUSED, -+ tree type ATTRIBUTE_UNUSED, -+ bool named ATTRIBUTE_UNUSED) -+ { -+ return 0; -+ } -+ -+ -+struct gcc_target targetm = TARGET_INITIALIZER; -+ -+/* -+ Table used to convert from register number in the assembler instructions and -+ the register numbers used in gcc. -+ */ -+const int avr32_function_arg_reglist[] = { -+ INTERNAL_REGNUM (12), -+ INTERNAL_REGNUM (11), -+ INTERNAL_REGNUM (10), -+ INTERNAL_REGNUM (9), -+ INTERNAL_REGNUM (8) -+}; -+ -+rtx avr32_compare_op0 = NULL_RTX; -+rtx avr32_compare_op1 = NULL_RTX; -+rtx avr32_compare_operator = NULL_RTX; -+rtx avr32_acc_cache = NULL_RTX; -+ -+/* -+ Returns nonzero if it is allowed to store a value of mode mode in hard -+ register number regno. -+ */ -+int -+avr32_hard_regno_mode_ok (int regnr, enum machine_mode mode) -+ { -+ /* We allow only float modes in the fp-registers */ -+ if (regnr >= FIRST_FP_REGNUM -+ && regnr <= LAST_FP_REGNUM && GET_MODE_CLASS (mode) != MODE_FLOAT) -+ { -+ return 0; -+ } -+ -+ switch (mode) -+ { -+ case DImode: /* long long */ -+ case DFmode: /* double */ -+ case SCmode: /* __complex__ float */ -+ case CSImode: /* __complex__ int */ -+ if (regnr < 4) -+ { /* long long int not supported in r12, sp, lr -+ or pc. */ -+ return 0; -+ } -+ else -+ { -+ if (regnr % 2) /* long long int has to be refered in even -+ registers. */ -+ return 0; -+ else -+ return 1; -+ } -+ case CDImode: /* __complex__ long long */ -+ case DCmode: /* __complex__ double */ -+ case TImode: /* 16 bytes */ -+ if (regnr < 7) -+ return 0; -+ else if (regnr % 2) -+ return 0; -+ else -+ return 1; -+ default: -+ return 1; -+ } -+ } -+ -+ -+int -+avr32_rnd_operands (rtx add, rtx shift) -+ { -+ if (GET_CODE (shift) == CONST_INT && -+ GET_CODE (add) == CONST_INT && INTVAL (shift) > 0) -+ { -+ if ((1 << (INTVAL (shift) - 1)) == INTVAL (add)) -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+ -+ -+int -+avr32_const_ok_for_constraint_p (HOST_WIDE_INT value, char c, const char *str) -+ { -+ switch (c) -+ { -+ case 'K': -+ case 'I': -+ { -+ HOST_WIDE_INT min_value = 0, max_value = 0; -+ char size_str[3]; -+ int const_size; -+ -+ size_str[0] = str[2]; -+ size_str[1] = str[3]; -+ size_str[2] = '\0'; -+ const_size = atoi (size_str); -+ -+ if (toupper (str[1]) == 'U') -+ { -+ min_value = 0; -+ max_value = (1 << const_size) - 1; -+ } -+ else if (toupper (str[1]) == 'S') -+ { -+ min_value = -(1 << (const_size - 1)); -+ max_value = (1 << (const_size - 1)) - 1; -+ } -+ -+ if (c == 'I') -+ { -+ value = -value; -+ } -+ -+ if (value >= min_value && value <= max_value) -+ { -+ return 1; -+ } -+ break; -+ } -+ case 'M': -+ return avr32_mask_upper_bits_operand (GEN_INT (value), VOIDmode); -+ } -+ -+ return 0; -+ } -+ -+ -+/*Compute mask of which floating-point registers needs saving upon -+ entry to this function*/ -+static unsigned long -+avr32_compute_save_fp_reg_mask (void) -+ { -+ unsigned long func_type = avr32_current_func_type (); -+ unsigned int save_reg_mask = 0; -+ unsigned int reg; -+ unsigned int max_reg = 7; -+ int save_all_call_used_regs = FALSE; -+ -+ /* This only applies for hardware floating-point implementation. */ -+ if (!TARGET_HARD_FLOAT) -+ return 0; -+ -+ if (IS_INTERRUPT (func_type)) -+ { -+ -+ /* Interrupt functions must not corrupt any registers, even call -+ clobbered ones. If this is a leaf function we can just examine the -+ registers used by the RTL, but otherwise we have to assume that -+ whatever function is called might clobber anything, and so we have -+ to save all the call-clobbered registers as well. */ -+ max_reg = 13; -+ save_all_call_used_regs = !current_function_is_leaf; -+ } -+ -+ /* All used registers used must be saved */ -+ for (reg = 0; reg <= max_reg; reg++) -+ if (regs_ever_live[INTERNAL_FP_REGNUM (reg)] -+ || (save_all_call_used_regs -+ && call_used_regs[INTERNAL_FP_REGNUM (reg)])) -+ save_reg_mask |= (1 << reg); -+ -+ return save_reg_mask; -+ } -+ -+/*Compute mask of registers which needs saving upon function entry */ -+static unsigned long -+avr32_compute_save_reg_mask (int push) -+ { -+ unsigned long func_type; -+ unsigned int save_reg_mask = 0; -+ unsigned int reg; -+ -+ func_type = avr32_current_func_type (); -+ -+ if (IS_INTERRUPT (func_type)) -+ { -+ unsigned int max_reg = 12; -+ -+ -+ /* Get the banking scheme for the interrupt */ -+ switch (func_type) -+ { -+ case AVR32_FT_ISR_FULL: -+ max_reg = 0; -+ break; -+ case AVR32_FT_ISR_HALF: -+ max_reg = 7; -+ break; -+ case AVR32_FT_ISR_NONE: -+ max_reg = 12; -+ break; -+ } -+ -+ /* Interrupt functions must not corrupt any registers, even call -+ clobbered ones. If this is a leaf function we can just examine the -+ registers used by the RTL, but otherwise we have to assume that -+ whatever function is called might clobber anything, and so we have -+ to save all the call-clobbered registers as well. */ -+ -+ /* Need not push the registers r8-r12 for AVR32A architectures, as this -+ is automatially done in hardware. We also do not have any shadow -+ registers. */ -+ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) -+ { -+ max_reg = 7; -+ func_type = AVR32_FT_ISR_NONE; -+ } -+ -+ /* All registers which are used and is not shadowed must be saved */ -+ for (reg = 0; reg <= max_reg; reg++) -+ if (regs_ever_live[INTERNAL_REGNUM (reg)] -+ || (!current_function_is_leaf -+ && call_used_regs[INTERNAL_REGNUM (reg)])) -+ save_reg_mask |= (1 << reg); -+ -+ /* Check LR */ -+ if ((regs_ever_live[LR_REGNUM] -+ || !current_function_is_leaf || frame_pointer_needed) -+ /* Only non-shadowed register models */ -+ && (func_type == AVR32_FT_ISR_NONE)) -+ save_reg_mask |= (1 << ASM_REGNUM (LR_REGNUM)); -+ -+ /* Make sure that the GOT register is pushed. */ -+ if (max_reg >= ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM) -+ && current_function_uses_pic_offset_table) -+ save_reg_mask |= (1 << ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM)); -+ -+ } -+ else -+ { -+ int use_pushm = optimize_size; -+ -+ /* In the normal case we only need to save those registers which are -+ call saved and which are used by this function. */ -+ for (reg = 0; reg <= 7; reg++) -+ if (regs_ever_live[INTERNAL_REGNUM (reg)] -+ && !call_used_regs[INTERNAL_REGNUM (reg)]) -+ save_reg_mask |= (1 << reg); -+ -+ /* Make sure that the GOT register is pushed. */ -+ if (current_function_uses_pic_offset_table) -+ save_reg_mask |= (1 << ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM)); -+ -+ -+ /* If we optimize for size and do not have anonymous arguments: use -+ popm/pushm always */ -+ if (use_pushm) -+ { -+ if ((save_reg_mask & (1 << 0)) -+ || (save_reg_mask & (1 << 1)) -+ || (save_reg_mask & (1 << 2)) || (save_reg_mask & (1 << 3))) -+ save_reg_mask |= 0xf; -+ -+ if ((save_reg_mask & (1 << 4)) -+ || (save_reg_mask & (1 << 5)) -+ || (save_reg_mask & (1 << 6)) || (save_reg_mask & (1 << 7))) -+ save_reg_mask |= 0xf0; -+ -+ if ((save_reg_mask & (1 << 8)) || (save_reg_mask & (1 << 9))) -+ save_reg_mask |= 0x300; -+ } -+ -+ -+ /* Check LR */ -+ if ((regs_ever_live[LR_REGNUM] -+ || !current_function_is_leaf -+ || (optimize_size -+ && save_reg_mask -+ && !current_function_calls_eh_return) || frame_pointer_needed)) -+ { -+ if (push -+ /* Never pop LR into PC for functions which -+ calls __builtin_eh_return, since we need to -+ fix the SP after the restoring of the registers -+ and before returning. */ -+ || current_function_calls_eh_return) -+ { -+ /* Push/Pop LR */ -+ save_reg_mask |= (1 << ASM_REGNUM (LR_REGNUM)); -+ } -+ else -+ { -+ /* Pop PC */ -+ save_reg_mask |= (1 << ASM_REGNUM (PC_REGNUM)); -+ } -+ } -+ } -+ -+ -+ /* Save registers so the exception handler can modify them. */ -+ if (current_function_calls_eh_return) -+ { -+ unsigned int i; -+ -+ for (i = 0;; i++) -+ { -+ reg = EH_RETURN_DATA_REGNO (i); -+ if (reg == INVALID_REGNUM) -+ break; -+ save_reg_mask |= 1 << ASM_REGNUM (reg); -+ } -+ } -+ -+ return save_reg_mask; -+ } -+ -+/*Compute total size in bytes of all saved registers */ -+static int -+avr32_get_reg_mask_size (int reg_mask) -+ { -+ int reg, size; -+ size = 0; -+ -+ for (reg = 0; reg <= 15; reg++) -+ if (reg_mask & (1 << reg)) -+ size += 4; -+ -+ return size; -+ } -+ -+/*Get a register from one of the registers which are saved onto the stack -+ upon function entry */ -+ -+static int -+avr32_get_saved_reg (int save_reg_mask) -+ { -+ unsigned int reg; -+ -+ /* Find the first register which is saved in the saved_reg_mask */ -+ for (reg = 0; reg <= 15; reg++) -+ if (save_reg_mask & (1 << reg)) -+ return reg; -+ -+ return -1; -+ } -+ -+/* Return 1 if it is possible to return using a single instruction. */ -+int -+avr32_use_return_insn (int iscond) -+ { -+ unsigned int func_type = avr32_current_func_type (); -+ unsigned long saved_int_regs; -+ unsigned long saved_fp_regs; -+ -+ /* Never use a return instruction before reload has run. */ -+ if (!reload_completed) -+ return 0; -+ -+ /* Must adjust the stack for vararg functions. */ -+ if (current_function_args_info.uses_anonymous_args) -+ return 0; -+ -+ /* If there a stack adjstment. */ -+ if (get_frame_size ()) -+ return 0; -+ -+ saved_int_regs = avr32_compute_save_reg_mask (TRUE); -+ saved_fp_regs = avr32_compute_save_fp_reg_mask (); -+ -+ /* Functions which have saved fp-regs on the stack can not be performed in -+ one instruction */ -+ if (saved_fp_regs) -+ return 0; -+ -+ /* Conditional returns can not be performed in one instruction if we need -+ to restore registers from the stack */ -+ if (iscond && saved_int_regs) -+ return 0; -+ -+ /* Conditional return can not be used for interrupt handlers. */ -+ if (iscond && IS_INTERRUPT (func_type)) -+ return 0; -+ -+ /* For interrupt handlers which needs to pop registers */ -+ if (saved_int_regs && IS_INTERRUPT (func_type)) -+ return 0; -+ -+ -+ /* If there are saved registers but the LR isn't saved, then we need two -+ instructions for the return. */ -+ if (saved_int_regs && !(saved_int_regs & (1 << ASM_REGNUM (LR_REGNUM)))) -+ return 0; -+ -+ -+ return 1; -+ } -+ -+ -+/*Generate some function prologue info in the assembly file*/ -+ -+void -+avr32_target_asm_function_prologue (FILE * f, HOST_WIDE_INT frame_size) -+ { -+ if (IS_NAKED (avr32_current_func_type ())) -+ fprintf (f, -+ "\t# Function is naked: Prologue and epilogue provided by programmer\n"); -+ -+ if (IS_INTERRUPT (avr32_current_func_type ())) -+ { -+ switch (avr32_current_func_type ()) -+ { -+ case AVR32_FT_ISR_FULL: -+ fprintf (f, -+ "\t# Interrupt Function: Fully shadowed register file\n"); -+ break; -+ case AVR32_FT_ISR_HALF: -+ fprintf (f, -+ "\t# Interrupt Function: Half shadowed register file\n"); -+ break; -+ default: -+ case AVR32_FT_ISR_NONE: -+ fprintf (f, "\t# Interrupt Function: No shadowed register file\n"); -+ break; -+ } -+ } -+ -+ -+ fprintf (f, "\t# args = %i, frame = %li, pretend = %i\n", -+ current_function_args_size, frame_size, -+ current_function_pretend_args_size); -+ -+ fprintf (f, "\t# frame_needed = %i, leaf_function = %i\n", -+ frame_pointer_needed, current_function_is_leaf); -+ -+ fprintf (f, "\t# uses_anonymous_args = %i\n", -+ current_function_args_info.uses_anonymous_args); -+ if (current_function_calls_eh_return) -+ fprintf (f, "\t# Calls __builtin_eh_return.\n"); -+ -+ } -+ -+ -+/* Generate and emit an insn that we will recognize as a pushm or stm. -+ Unfortunately, since this insn does not reflect very well the actual -+ semantics of the operation, we need to annotate the insn for the benefit -+ of DWARF2 frame unwind information. */ -+ -+int avr32_convert_to_reglist16 (int reglist8_vect); -+ -+static rtx -+emit_multi_reg_push (int reglist, int usePUSHM) -+ { -+ rtx insn; -+ rtx dwarf; -+ rtx tmp; -+ rtx reg; -+ int i; -+ int nr_regs; -+ int index = 0; -+ -+ if (usePUSHM) -+ { -+ insn = emit_insn (gen_pushm (gen_rtx_CONST_INT (SImode, reglist))); -+ reglist = avr32_convert_to_reglist16 (reglist); -+ } -+ else -+ { -+ insn = emit_insn (gen_stm (stack_pointer_rtx, -+ gen_rtx_CONST_INT (SImode, reglist), -+ gen_rtx_CONST_INT (SImode, 1))); -+ } -+ -+ nr_regs = avr32_get_reg_mask_size (reglist) / 4; -+ dwarf = gen_rtx_SEQUENCE (VOIDmode, rtvec_alloc (nr_regs + 1)); -+ -+ for (i = 15; i >= 0; i--) -+ { -+ if (reglist & (1 << i)) -+ { -+ reg = gen_rtx_REG (SImode, INTERNAL_REGNUM (i)); -+ tmp = gen_rtx_SET (VOIDmode, -+ gen_rtx_MEM (SImode, -+ plus_constant (stack_pointer_rtx, -+ 4 * index)), reg); -+ RTX_FRAME_RELATED_P (tmp) = 1; -+ XVECEXP (dwarf, 0, 1 + index++) = tmp; -+ } -+ } -+ -+ tmp = gen_rtx_SET (SImode, -+ stack_pointer_rtx, -+ gen_rtx_PLUS (SImode, -+ stack_pointer_rtx, -+ GEN_INT (-4 * nr_regs))); -+ RTX_FRAME_RELATED_P (tmp) = 1; -+ XVECEXP (dwarf, 0, 0) = tmp; -+ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, dwarf, -+ REG_NOTES (insn)); -+ return insn; -+ } -+ -+ -+static rtx -+emit_multi_fp_reg_push (int reglist) -+ { -+ rtx insn; -+ rtx dwarf; -+ rtx tmp; -+ rtx reg; -+ int i; -+ int nr_regs; -+ int index = 0; -+ -+ insn = emit_insn (gen_stm_fp (stack_pointer_rtx, -+ gen_rtx_CONST_INT (SImode, reglist), -+ gen_rtx_CONST_INT (SImode, 1))); -+ -+ nr_regs = avr32_get_reg_mask_size (reglist) / 4; -+ dwarf = gen_rtx_SEQUENCE (VOIDmode, rtvec_alloc (nr_regs + 1)); -+ -+ for (i = 15; i >= 0; i--) -+ { -+ if (reglist & (1 << i)) -+ { -+ reg = gen_rtx_REG (SImode, INTERNAL_FP_REGNUM (i)); -+ tmp = gen_rtx_SET (VOIDmode, -+ gen_rtx_MEM (SImode, -+ plus_constant (stack_pointer_rtx, -+ 4 * index)), reg); -+ RTX_FRAME_RELATED_P (tmp) = 1; -+ XVECEXP (dwarf, 0, 1 + index++) = tmp; -+ } -+ } -+ -+ tmp = gen_rtx_SET (SImode, -+ stack_pointer_rtx, -+ gen_rtx_PLUS (SImode, -+ stack_pointer_rtx, -+ GEN_INT (-4 * nr_regs))); -+ RTX_FRAME_RELATED_P (tmp) = 1; -+ XVECEXP (dwarf, 0, 0) = tmp; -+ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, dwarf, -+ REG_NOTES (insn)); -+ return insn; -+ } -+ -+rtx -+avr32_gen_load_multiple (rtx * regs, int count, rtx from, -+ int write_back, int in_struct_p, int scalar_p) -+ { -+ -+ rtx result; -+ int i = 0, j; -+ -+ result = -+ gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (count + (write_back ? 1 : 0))); -+ -+ if (write_back) -+ { -+ XVECEXP (result, 0, 0) -+ = gen_rtx_SET (GET_MODE (from), from, -+ plus_constant (from, count * 4)); -+ i = 1; -+ count++; -+ } -+ -+ -+ for (j = 0; i < count; i++, j++) -+ { -+ rtx unspec; -+ rtx mem = gen_rtx_MEM (SImode, plus_constant (from, j * 4)); -+ MEM_IN_STRUCT_P (mem) = in_struct_p; -+ MEM_SCALAR_P (mem) = scalar_p; -+ unspec = gen_rtx_UNSPEC (VOIDmode, gen_rtvec (1, mem), UNSPEC_LDM); -+ XVECEXP (result, 0, i) = gen_rtx_SET (VOIDmode, regs[j], unspec); -+ } -+ -+ return result; -+ } -+ -+ -+rtx -+avr32_gen_store_multiple (rtx * regs, int count, rtx to, -+ int in_struct_p, int scalar_p) -+ { -+ rtx result; -+ int i = 0, j; -+ -+ result = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (count)); -+ -+ for (j = 0; i < count; i++, j++) -+ { -+ rtx mem = gen_rtx_MEM (SImode, plus_constant (to, j * 4)); -+ MEM_IN_STRUCT_P (mem) = in_struct_p; -+ MEM_SCALAR_P (mem) = scalar_p; -+ XVECEXP (result, 0, i) -+ = gen_rtx_SET (VOIDmode, mem, -+ gen_rtx_UNSPEC (VOIDmode, -+ gen_rtvec (1, regs[j]), -+ UNSPEC_STORE_MULTIPLE)); -+ } -+ -+ return result; -+ } -+ -+ -+/* Move a block of memory if it is word aligned or we support unaligned -+ word memory accesses. The size must be maximum 64 bytes. */ -+ -+int -+avr32_gen_movmemsi (rtx * operands) -+ { -+ HOST_WIDE_INT bytes_to_go; -+ rtx src, dst; -+ rtx st_src, st_dst; -+ int ptr_offset = 0; -+ int block_size; -+ int dst_in_struct_p, src_in_struct_p; -+ int dst_scalar_p, src_scalar_p; -+ int unaligned; -+ -+ if (GET_CODE (operands[2]) != CONST_INT -+ || GET_CODE (operands[3]) != CONST_INT -+ || INTVAL (operands[2]) > 64 -+ || ((INTVAL (operands[3]) & 3) && !TARGET_UNALIGNED_WORD)) -+ return 0; -+ -+ unaligned = (INTVAL (operands[3]) & 3) != 0; -+ -+ block_size = 4; -+ -+ st_dst = XEXP (operands[0], 0); -+ st_src = XEXP (operands[1], 0); -+ -+ dst_in_struct_p = MEM_IN_STRUCT_P (operands[0]); -+ dst_scalar_p = MEM_SCALAR_P (operands[0]); -+ src_in_struct_p = MEM_IN_STRUCT_P (operands[1]); -+ src_scalar_p = MEM_SCALAR_P (operands[1]); -+ -+ dst = copy_to_mode_reg (SImode, st_dst); -+ src = copy_to_mode_reg (SImode, st_src); -+ -+ bytes_to_go = INTVAL (operands[2]); -+ -+ while (bytes_to_go) -+ { -+ enum machine_mode move_mode; -+ /* (Seems to be a problem with reloads for the movti pattern so this is -+ disabled until that problem is resolved) -+ UPDATE: Problem seems to be solved now.... */ -+ if (bytes_to_go >= GET_MODE_SIZE (TImode) && !unaligned -+ /* Do not emit ldm/stm for UC3 as ld.d/st.d is more optimal. */ -+ && avr32_arch->arch_type != ARCH_TYPE_AVR32_UC) -+ move_mode = TImode; -+ else if ((bytes_to_go >= GET_MODE_SIZE (DImode)) && !unaligned) -+ move_mode = DImode; -+ else if (bytes_to_go >= GET_MODE_SIZE (SImode)) -+ move_mode = SImode; -+ else -+ move_mode = QImode; -+ -+ { -+ rtx dst_mem = gen_rtx_MEM (move_mode, -+ gen_rtx_PLUS (SImode, dst, -+ GEN_INT (ptr_offset))); -+ rtx src_mem = gen_rtx_MEM (move_mode, -+ gen_rtx_PLUS (SImode, src, -+ GEN_INT (ptr_offset))); -+ ptr_offset += GET_MODE_SIZE (move_mode); -+ bytes_to_go -= GET_MODE_SIZE (move_mode); -+ -+ MEM_IN_STRUCT_P (dst_mem) = dst_in_struct_p; -+ MEM_SCALAR_P (dst_mem) = dst_scalar_p; -+ -+ MEM_IN_STRUCT_P (src_mem) = src_in_struct_p; -+ MEM_SCALAR_P (src_mem) = src_scalar_p; -+ emit_move_insn (dst_mem, src_mem); -+ -+ } -+ } -+ -+ return 1; -+ } -+ -+ -+ -+/*Expand the prologue instruction*/ -+void -+avr32_expand_prologue (void) -+ { -+ rtx insn, dwarf; -+ unsigned long saved_reg_mask, saved_fp_reg_mask; -+ int reglist8 = 0; -+ -+ /* Naked functions does not have a prologue */ -+ if (IS_NAKED (avr32_current_func_type ())) -+ return; -+ -+ saved_reg_mask = avr32_compute_save_reg_mask (TRUE); -+ -+ if (saved_reg_mask) -+ { -+ /* Must push used registers */ -+ -+ /* Should we use POPM or LDM? */ -+ int usePUSHM = TRUE; -+ reglist8 = 0; -+ if (((saved_reg_mask & (1 << 0)) || -+ (saved_reg_mask & (1 << 1)) || -+ (saved_reg_mask & (1 << 2)) || (saved_reg_mask & (1 << 3)))) -+ { -+ /* One of R0-R3 should at least be pushed */ -+ if (((saved_reg_mask & (1 << 0)) && -+ (saved_reg_mask & (1 << 1)) && -+ (saved_reg_mask & (1 << 2)) && (saved_reg_mask & (1 << 3)))) -+ { -+ /* All should be pushed */ -+ reglist8 |= 0x01; -+ } -+ else -+ { -+ usePUSHM = FALSE; -+ } -+ } -+ -+ if (((saved_reg_mask & (1 << 4)) || -+ (saved_reg_mask & (1 << 5)) || -+ (saved_reg_mask & (1 << 6)) || (saved_reg_mask & (1 << 7)))) -+ { -+ /* One of R4-R7 should at least be pushed */ -+ if (((saved_reg_mask & (1 << 4)) && -+ (saved_reg_mask & (1 << 5)) && -+ (saved_reg_mask & (1 << 6)) && (saved_reg_mask & (1 << 7)))) -+ { -+ if (usePUSHM) -+ /* All should be pushed */ -+ reglist8 |= 0x02; -+ } -+ else -+ { -+ usePUSHM = FALSE; -+ } -+ } -+ -+ if (((saved_reg_mask & (1 << 8)) || (saved_reg_mask & (1 << 9)))) -+ { -+ /* One of R8-R9 should at least be pushed */ -+ if (((saved_reg_mask & (1 << 8)) && (saved_reg_mask & (1 << 9)))) -+ { -+ if (usePUSHM) -+ /* All should be pushed */ -+ reglist8 |= 0x04; -+ } -+ else -+ { -+ usePUSHM = FALSE; -+ } -+ } -+ -+ if (saved_reg_mask & (1 << 10)) -+ reglist8 |= 0x08; -+ -+ if (saved_reg_mask & (1 << 11)) -+ reglist8 |= 0x10; -+ -+ if (saved_reg_mask & (1 << 12)) -+ reglist8 |= 0x20; -+ -+ if (saved_reg_mask & (1 << ASM_REGNUM (LR_REGNUM))) -+ { -+ /* Push LR */ -+ reglist8 |= 0x40; -+ } -+ -+ if (usePUSHM) -+ { -+ insn = emit_multi_reg_push (reglist8, TRUE); -+ } -+ else -+ { -+ insn = emit_multi_reg_push (saved_reg_mask, FALSE); -+ } -+ RTX_FRAME_RELATED_P (insn) = 1; -+ -+ /* Prevent this instruction from being scheduled after any other -+ instructions. */ -+ emit_insn (gen_blockage ()); -+ } -+ -+ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); -+ if (saved_fp_reg_mask) -+ { -+ insn = emit_multi_fp_reg_push (saved_fp_reg_mask); -+ RTX_FRAME_RELATED_P (insn) = 1; -+ -+ /* Prevent this instruction from being scheduled after any other -+ instructions. */ -+ emit_insn (gen_blockage ()); -+ } -+ -+ /* Set frame pointer */ -+ if (frame_pointer_needed) -+ { -+ insn = emit_move_insn (frame_pointer_rtx, stack_pointer_rtx); -+ RTX_FRAME_RELATED_P (insn) = 1; -+ } -+ -+ if (get_frame_size () > 0) -+ { -+ if (avr32_const_ok_for_constraint_p (get_frame_size (), 'K', "Ks21")) -+ { -+ insn = emit_insn (gen_rtx_SET (SImode, -+ stack_pointer_rtx, -+ gen_rtx_PLUS (SImode, -+ stack_pointer_rtx, -+ gen_rtx_CONST_INT -+ (SImode, -+ -get_frame_size -+ ())))); -+ RTX_FRAME_RELATED_P (insn) = 1; -+ } -+ else -+ { -+ /* Immediate is larger than k21 We must either check if we can use -+ one of the pushed reegisters as temporary storage or we must -+ make us a temp register by pushing a register to the stack. */ -+ rtx temp_reg, const_pool_entry, insn; -+ if (saved_reg_mask) -+ { -+ temp_reg = -+ gen_rtx_REG (SImode, -+ INTERNAL_REGNUM (avr32_get_saved_reg -+ (saved_reg_mask))); -+ } -+ else -+ { -+ temp_reg = gen_rtx_REG (SImode, INTERNAL_REGNUM (7)); -+ emit_move_insn (gen_rtx_MEM -+ (SImode, -+ gen_rtx_PRE_DEC (SImode, stack_pointer_rtx)), -+ temp_reg); -+ } -+ -+ const_pool_entry = -+ force_const_mem (SImode, -+ gen_rtx_CONST_INT (SImode, get_frame_size ())); -+ emit_move_insn (temp_reg, const_pool_entry); -+ -+ insn = emit_insn (gen_rtx_SET (SImode, -+ stack_pointer_rtx, -+ gen_rtx_MINUS (SImode, -+ stack_pointer_rtx, -+ temp_reg))); -+ -+ dwarf = gen_rtx_SET (VOIDmode, stack_pointer_rtx, -+ gen_rtx_PLUS (SImode, stack_pointer_rtx, -+ GEN_INT (-get_frame_size ()))); -+ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, -+ dwarf, REG_NOTES (insn)); -+ RTX_FRAME_RELATED_P (insn) = 1; -+ -+ if (!saved_reg_mask) -+ { -+ insn = -+ emit_move_insn (temp_reg, -+ gen_rtx_MEM (SImode, -+ gen_rtx_POST_INC (SImode, -+ gen_rtx_REG -+ (SImode, -+ 13)))); -+ } -+ -+ /* Mark the temp register as dead */ -+ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_DEAD, temp_reg, -+ REG_NOTES (insn)); -+ -+ -+ } -+ -+ /* Prevent the the stack adjustment to be scheduled after any -+ instructions using the frame pointer. */ -+ emit_insn (gen_blockage ()); -+ } -+ -+ /* Load GOT */ -+ if (flag_pic) -+ { -+ avr32_load_pic_register (); -+ -+ /* gcc does not know that load or call instructions might use the pic -+ register so it might schedule these instructions before the loading -+ of the pic register. To avoid this emit a barrier for now. TODO! -+ Find out a better way to let gcc know which instructions might use -+ the pic register. */ -+ emit_insn (gen_blockage ()); -+ } -+ return; -+ } -+ -+void -+avr32_set_return_address (rtx source, rtx scratch) -+ { -+ rtx addr; -+ unsigned long saved_regs; -+ -+ saved_regs = avr32_compute_save_reg_mask (TRUE); -+ -+ if (!(saved_regs & (1 << ASM_REGNUM (LR_REGNUM)))) -+ emit_move_insn (gen_rtx_REG (Pmode, LR_REGNUM), source); -+ else -+ { -+ if (frame_pointer_needed) -+ addr = gen_rtx_REG (Pmode, FRAME_POINTER_REGNUM); -+ else -+ if (avr32_const_ok_for_constraint_p (get_frame_size (), 'K', "Ks16")) -+ { -+ addr = plus_constant (stack_pointer_rtx, get_frame_size ()); -+ } -+ else -+ { -+ emit_insn (gen_movsi (scratch, GEN_INT (get_frame_size ()))); -+ addr = scratch; -+ } -+ emit_move_insn (gen_rtx_MEM (Pmode, addr), source); -+ } -+ } -+ -+ -+ -+/* Return the length of INSN. LENGTH is the initial length computed by -+ attributes in the machine-description file. */ -+ -+int -+avr32_adjust_insn_length (rtx insn ATTRIBUTE_UNUSED, -+ int length ATTRIBUTE_UNUSED) -+ { -+ return length; -+ } -+ -+void -+avr32_output_return_instruction (int single_ret_inst ATTRIBUTE_UNUSED, -+ int iscond ATTRIBUTE_UNUSED, -+ rtx cond ATTRIBUTE_UNUSED, rtx r12_imm) -+ { -+ -+ unsigned long saved_reg_mask, saved_fp_reg_mask; -+ int insert_ret = TRUE; -+ int reglist8 = 0; -+ int stack_adjustment = get_frame_size (); -+ unsigned int func_type = avr32_current_func_type (); -+ FILE *f = asm_out_file; -+ -+ /* Naked functions does not have an epilogue */ -+ if (IS_NAKED (func_type)) -+ return; -+ -+ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); -+ -+ saved_reg_mask = avr32_compute_save_reg_mask (FALSE); -+ -+ /* Reset frame pointer */ -+ if (stack_adjustment > 0) -+ { -+ if (avr32_const_ok_for_constraint_p (stack_adjustment, 'I', "Is21")) -+ { -+ fprintf (f, "\tsub\tsp, %i # Reset Frame Pointer\n", -+ -stack_adjustment); -+ } -+ else -+ { -+ /* TODO! Is it safe to use r8 as scratch?? */ -+ fprintf (f, "\tmov\tr8, lo(%i) # Reset Frame Pointer\n", -+ -stack_adjustment); -+ fprintf (f, "\torh\tr8, hi(%i) # Reset Frame Pointer\n", -+ -stack_adjustment); -+ fprintf (f, "\tadd\tsp, r8 # Reset Frame Pointer\n"); -+ } -+ } -+ -+ if (saved_fp_reg_mask) -+ { -+ char reglist[64]; /* 64 bytes should be enough... */ -+ avr32_make_fp_reglist_w (saved_fp_reg_mask, (char *) reglist); -+ fprintf (f, "\tldcm.w\tcp0, sp++, %s\n", reglist); -+ if (saved_fp_reg_mask & ~0xff) -+ { -+ saved_fp_reg_mask &= ~0xff; -+ avr32_make_fp_reglist_d (saved_fp_reg_mask, (char *) reglist); -+ fprintf (f, "\tldcm.d\tcp0, sp++, %s\n", reglist); -+ } -+ } -+ -+ if (saved_reg_mask) -+ { -+ /* Must pop used registers */ -+ -+ /* Should we use POPM or LDM? */ -+ int usePOPM = TRUE; -+ if (((saved_reg_mask & (1 << 0)) || -+ (saved_reg_mask & (1 << 1)) || -+ (saved_reg_mask & (1 << 2)) || (saved_reg_mask & (1 << 3)))) -+ { -+ /* One of R0-R3 should at least be popped */ -+ if (((saved_reg_mask & (1 << 0)) && -+ (saved_reg_mask & (1 << 1)) && -+ (saved_reg_mask & (1 << 2)) && (saved_reg_mask & (1 << 3)))) -+ { -+ /* All should be popped */ -+ reglist8 |= 0x01; -+ } -+ else -+ { -+ usePOPM = FALSE; -+ } -+ } -+ -+ if (((saved_reg_mask & (1 << 4)) || -+ (saved_reg_mask & (1 << 5)) || -+ (saved_reg_mask & (1 << 6)) || (saved_reg_mask & (1 << 7)))) -+ { -+ /* One of R0-R3 should at least be popped */ -+ if (((saved_reg_mask & (1 << 4)) && -+ (saved_reg_mask & (1 << 5)) && -+ (saved_reg_mask & (1 << 6)) && (saved_reg_mask & (1 << 7)))) -+ { -+ if (usePOPM) -+ /* All should be popped */ -+ reglist8 |= 0x02; -+ } -+ else -+ { -+ usePOPM = FALSE; -+ } -+ } -+ -+ if (((saved_reg_mask & (1 << 8)) || (saved_reg_mask & (1 << 9)))) -+ { -+ /* One of R8-R9 should at least be pushed */ -+ if (((saved_reg_mask & (1 << 8)) && (saved_reg_mask & (1 << 9)))) -+ { -+ if (usePOPM) -+ /* All should be pushed */ -+ reglist8 |= 0x04; -+ } -+ else -+ { -+ usePOPM = FALSE; -+ } -+ } -+ -+ if (saved_reg_mask & (1 << 10)) -+ reglist8 |= 0x08; -+ -+ if (saved_reg_mask & (1 << 11)) -+ reglist8 |= 0x10; -+ -+ if (saved_reg_mask & (1 << 12)) -+ reglist8 |= 0x20; -+ -+ if (saved_reg_mask & (1 << ASM_REGNUM (LR_REGNUM))) -+ /* Pop LR */ -+ reglist8 |= 0x40; -+ -+ if (saved_reg_mask & (1 << ASM_REGNUM (PC_REGNUM))) -+ /* Pop LR into PC. */ -+ reglist8 |= 0x80; -+ -+ if (usePOPM) -+ { -+ char reglist[64]; /* 64 bytes should be enough... */ -+ avr32_make_reglist8 (reglist8, (char *) reglist); -+ -+ if (reglist8 & 0x80) -+ /* This instruction is also a return */ -+ insert_ret = FALSE; -+ -+ if (r12_imm && !insert_ret) -+ fprintf (f, "\tpopm\t%s, r12=%li\n", reglist, INTVAL (r12_imm)); -+ else -+ fprintf (f, "\tpopm\t%s\n", reglist); -+ -+ } -+ else -+ { -+ char reglist[64]; /* 64 bytes should be enough... */ -+ avr32_make_reglist16 (saved_reg_mask, (char *) reglist); -+ if (saved_reg_mask & (1 << ASM_REGNUM (PC_REGNUM))) -+ /* This instruction is also a return */ -+ insert_ret = FALSE; -+ -+ if (r12_imm && !insert_ret) -+ fprintf (f, "\tldm\tsp++, %s, r12=%li\n", reglist, -+ INTVAL (r12_imm)); -+ else -+ fprintf (f, "\tldm\tsp++, %s\n", reglist); -+ -+ } -+ -+ } -+ -+ /* Stack adjustment for exception handler. */ -+ if (current_function_calls_eh_return) -+ fprintf (f, "\tadd\tsp, r%d\n", ASM_REGNUM (EH_RETURN_STACKADJ_REGNO)); -+ -+ -+ if (IS_INTERRUPT (func_type)) -+ { -+ fprintf (f, "\trete\n"); -+ } -+ else if (insert_ret) -+ { -+ if (r12_imm) -+ fprintf (f, "\tretal\t%li\n", INTVAL (r12_imm)); -+ else -+ fprintf (f, "\tretal\tr12\n"); -+ } -+ } -+ -+/* Function for converting a fp-register mask to a -+ reglistCPD8 register list string. */ -+void -+avr32_make_fp_reglist_d (int reglist_mask, char *reglist_string) -+ { -+ int i; -+ -+ /* Make sure reglist_string is empty */ -+ reglist_string[0] = '\0'; -+ -+ for (i = 0; i < NUM_FP_REGS; i += 2) -+ { -+ if (reglist_mask & (1 << i)) -+ { -+ strlen (reglist_string) ? -+ sprintf (reglist_string, "%s, %s-%s", reglist_string, -+ reg_names[INTERNAL_FP_REGNUM (i)], -+ reg_names[INTERNAL_FP_REGNUM (i + 1)]) : -+ sprintf (reglist_string, "%s-%s", -+ reg_names[INTERNAL_FP_REGNUM (i)], -+ reg_names[INTERNAL_FP_REGNUM (i + 1)]); -+ } -+ } -+ } -+ -+/* Function for converting a fp-register mask to a -+ reglistCP8 register list string. */ -+void -+avr32_make_fp_reglist_w (int reglist_mask, char *reglist_string) -+ { -+ int i; -+ -+ /* Make sure reglist_string is empty */ -+ reglist_string[0] = '\0'; -+ -+ for (i = 0; i < NUM_FP_REGS; ++i) -+ { -+ if (reglist_mask & (1 << i)) -+ { -+ strlen (reglist_string) ? -+ sprintf (reglist_string, "%s, %s", reglist_string, -+ reg_names[INTERNAL_FP_REGNUM (i)]) : -+ sprintf (reglist_string, "%s", reg_names[INTERNAL_FP_REGNUM (i)]); -+ } -+ } -+ } -+ -+void -+avr32_make_reglist16 (int reglist16_vect, char *reglist16_string) -+ { -+ int i; -+ -+ /* Make sure reglist16_string is empty */ -+ reglist16_string[0] = '\0'; -+ -+ for (i = 0; i < 16; ++i) -+ { -+ if (reglist16_vect & (1 << i)) -+ { -+ strlen (reglist16_string) ? -+ sprintf (reglist16_string, "%s, %s", reglist16_string, -+ reg_names[INTERNAL_REGNUM (i)]) : -+ sprintf (reglist16_string, "%s", reg_names[INTERNAL_REGNUM (i)]); -+ } -+ } -+ } -+ -+int -+avr32_convert_to_reglist16 (int reglist8_vect) -+ { -+ int reglist16_vect = 0; -+ if (reglist8_vect & 0x1) -+ reglist16_vect |= 0xF; -+ if (reglist8_vect & 0x2) -+ reglist16_vect |= 0xF0; -+ if (reglist8_vect & 0x4) -+ reglist16_vect |= 0x300; -+ if (reglist8_vect & 0x8) -+ reglist16_vect |= 0x400; -+ if (reglist8_vect & 0x10) -+ reglist16_vect |= 0x800; -+ if (reglist8_vect & 0x20) -+ reglist16_vect |= 0x1000; -+ if (reglist8_vect & 0x40) -+ reglist16_vect |= 0x4000; -+ if (reglist8_vect & 0x80) -+ reglist16_vect |= 0x8000; -+ -+ return reglist16_vect; -+ } -+ -+void -+avr32_make_reglist8 (int reglist8_vect, char *reglist8_string) -+ { -+ /* Make sure reglist8_string is empty */ -+ reglist8_string[0] = '\0'; -+ -+ if (reglist8_vect & 0x1) -+ sprintf (reglist8_string, "r0-r3"); -+ if (reglist8_vect & 0x2) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r4-r7", -+ reglist8_string) : -+ sprintf (reglist8_string, "r4-r7"); -+ if (reglist8_vect & 0x4) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r8-r9", -+ reglist8_string) : -+ sprintf (reglist8_string, "r8-r9"); -+ if (reglist8_vect & 0x8) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r10", -+ reglist8_string) : -+ sprintf (reglist8_string, "r10"); -+ if (reglist8_vect & 0x10) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r11", -+ reglist8_string) : -+ sprintf (reglist8_string, "r11"); -+ if (reglist8_vect & 0x20) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r12", -+ reglist8_string) : -+ sprintf (reglist8_string, "r12"); -+ if (reglist8_vect & 0x40) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, lr", -+ reglist8_string) : -+ sprintf (reglist8_string, "lr"); -+ if (reglist8_vect & 0x80) -+ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, pc", -+ reglist8_string) : -+ sprintf (reglist8_string, "pc"); -+ } -+ -+int -+avr32_eh_return_data_regno (int n) -+ { -+ if (n >= 0 && n <= 3) -+ return 8 + n; -+ else -+ return INVALID_REGNUM; -+ } -+ -+/* Compute the distance from register FROM to register TO. -+ These can be the arg pointer, the frame pointer or -+ the stack pointer. -+ Typical stack layout looks like this: -+ -+ old stack pointer -> | | -+ ---- -+ | | \ -+ | | saved arguments for -+ | | vararg functions -+ arg_pointer -> | | / -+ -- -+ | | \ -+ | | call saved -+ | | registers -+ | | / -+ frame ptr -> -- -+ | | \ -+ | | local -+ | | variables -+ stack ptr --> | | / -+ -- -+ | | \ -+ | | outgoing -+ | | arguments -+ | | / -+ -- -+ -+ For a given funciton some or all of these stack compomnents -+ may not be needed, giving rise to the possibility of -+ eliminating some of the registers. -+ -+ The values returned by this function must reflect the behaviour -+ of avr32_expand_prologue() and avr32_compute_save_reg_mask(). -+ -+ The sign of the number returned reflects the direction of stack -+ growth, so the values are positive for all eliminations except -+ from the soft frame pointer to the hard frame pointer. */ -+ -+ -+int -+avr32_initial_elimination_offset (int from, int to) -+ { -+ int i; -+ int call_saved_regs = 0; -+ unsigned long saved_reg_mask, saved_fp_reg_mask; -+ unsigned int local_vars = get_frame_size (); -+ -+ saved_reg_mask = avr32_compute_save_reg_mask (TRUE); -+ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); -+ -+ for (i = 0; i < 16; ++i) -+ { -+ if (saved_reg_mask & (1 << i)) -+ call_saved_regs += 4; -+ } -+ -+ for (i = 0; i < NUM_FP_REGS; ++i) -+ { -+ if (saved_fp_reg_mask & (1 << i)) -+ call_saved_regs += 4; -+ } -+ -+ switch (from) -+ { -+ case ARG_POINTER_REGNUM: -+ switch (to) -+ { -+ case STACK_POINTER_REGNUM: -+ return call_saved_regs + local_vars; -+ case FRAME_POINTER_REGNUM: -+ return call_saved_regs; -+ default: -+ abort (); -+ } -+ case FRAME_POINTER_REGNUM: -+ switch (to) -+ { -+ case STACK_POINTER_REGNUM: -+ return local_vars; -+ default: -+ abort (); -+ } -+ default: -+ abort (); -+ } -+ } -+ -+ -+/* -+ Returns a rtx used when passing the next argument to a function. -+ avr32_init_cumulative_args() and avr32_function_arg_advance() sets witch -+ register to use. -+ */ -+rtx -+avr32_function_arg (CUMULATIVE_ARGS * cum, enum machine_mode mode, -+ tree type, int named) -+ { -+ int index = -1; -+ -+ HOST_WIDE_INT arg_size, arg_rsize; -+ if (type) -+ { -+ arg_size = int_size_in_bytes (type); -+ } -+ else -+ { -+ arg_size = GET_MODE_SIZE (mode); -+ } -+ arg_rsize = PUSH_ROUNDING (arg_size); -+ -+ /* -+ The last time this macro is called, it is called with mode == VOIDmode, -+ and its result is passed to the call or call_value pattern as operands 2 -+ and 3 respectively. */ -+ if (mode == VOIDmode) -+ { -+ return gen_rtx_CONST_INT (SImode, 22); /* ToDo: fixme. */ -+ } -+ -+ if ((*targetm.calls.must_pass_in_stack) (mode, type) || !named) -+ { -+ return NULL_RTX; -+ } -+ -+ if (arg_rsize == 8) -+ { -+ /* use r11:r10 or r9:r8. */ -+ if (!(GET_USED_INDEX (cum, 1) || GET_USED_INDEX (cum, 2))) -+ index = 1; -+ else if (!(GET_USED_INDEX (cum, 3) || GET_USED_INDEX (cum, 4))) -+ index = 3; -+ else -+ index = -1; -+ } -+ else if (arg_rsize == 4) -+ { /* Use first available register */ -+ index = 0; -+ while (index <= LAST_CUM_REG_INDEX && GET_USED_INDEX (cum, index)) -+ index++; -+ if (index > LAST_CUM_REG_INDEX) -+ index = -1; -+ } -+ -+ SET_REG_INDEX (cum, index); -+ -+ if (GET_REG_INDEX (cum) >= 0) -+ return gen_rtx_REG (mode, -+ avr32_function_arg_reglist[GET_REG_INDEX (cum)]); -+ -+ return NULL_RTX; -+ } -+ -+/* -+ Set the register used for passing the first argument to a function. -+ */ -+void -+avr32_init_cumulative_args (CUMULATIVE_ARGS * cum, tree fntype, -+ rtx libname ATTRIBUTE_UNUSED, -+ tree fndecl ATTRIBUTE_UNUSED) -+ { -+ /* Set all registers as unused. */ -+ SET_INDEXES_UNUSED (cum); -+ -+ /* Reset uses_anonymous_args */ -+ cum->uses_anonymous_args = 0; -+ -+ /* Reset size of stack pushed arguments */ -+ cum->stack_pushed_args_size = 0; -+ } -+ -+/* -+ Set register used for passing the next argument to a function. Only the -+ Scratch Registers are used. -+ -+ number name -+ 15 r15 PC -+ 14 r14 LR -+ 13 r13 _SP_________ -+ FIRST_CUM_REG 12 r12 _||_ -+ 10 r11 || -+ 11 r10 _||_ Scratch Registers -+ 8 r9 || -+ LAST_SCRATCH_REG 9 r8 _\/_________ -+ 6 r7 /\ -+ 7 r6 || -+ 4 r5 || -+ 5 r4 || -+ 2 r3 || -+ 3 r2 || -+ 0 r1 || -+ 1 r0 _||_________ -+ -+ */ -+void -+avr32_function_arg_advance (CUMULATIVE_ARGS * cum, enum machine_mode mode, -+ tree type, int named ATTRIBUTE_UNUSED) -+ { -+ HOST_WIDE_INT arg_size, arg_rsize; -+ -+ if (type) -+ { -+ arg_size = int_size_in_bytes (type); -+ } -+ else -+ { -+ arg_size = GET_MODE_SIZE (mode); -+ } -+ arg_rsize = PUSH_ROUNDING (arg_size); -+ -+ /* It the argument had to be passed in stack, no register is used. */ -+ if ((*targetm.calls.must_pass_in_stack) (mode, type)) -+ { -+ cum->stack_pushed_args_size += PUSH_ROUNDING (int_size_in_bytes (type)); -+ return; -+ } -+ -+ /* Mark the used registers as "used". */ -+ if (GET_REG_INDEX (cum) >= 0) -+ { -+ SET_USED_INDEX (cum, GET_REG_INDEX (cum)); -+ if (arg_rsize == 8) -+ { -+ SET_USED_INDEX (cum, (GET_REG_INDEX (cum) + 1)); -+ } -+ } -+ else -+ { -+ /* Had to use stack */ -+ cum->stack_pushed_args_size += arg_rsize; -+ } -+ } -+ -+/* -+ Defines witch direction to go to find the next register to use if the -+ argument is larger then one register or for arguments shorter than an -+ int which is not promoted, such as the last part of structures with -+ size not a multiple of 4. */ -+enum direction -+avr32_function_arg_padding (enum machine_mode mode ATTRIBUTE_UNUSED, -+ tree type) -+ { -+ /* Pad upward for all aggregates except byte and halfword sized aggregates -+ which can be passed in registers. */ -+ if (type -+ && AGGREGATE_TYPE_P (type) -+ && (int_size_in_bytes (type) != 1) -+ && !((int_size_in_bytes (type) == 2) -+ && TYPE_ALIGN_UNIT (type) >= 2) -+ && (int_size_in_bytes (type) & 0x3)) -+ { -+ return upward; -+ } -+ -+ return downward; -+ } -+ -+/* -+ Return a rtx used for the return value from a function call. -+ */ -+rtx -+avr32_function_value (tree type, tree func, bool outgoing ATTRIBUTE_UNUSED) -+ { -+ if (avr32_return_in_memory (type, func)) -+ return NULL_RTX; -+ -+ if (int_size_in_bytes (type) <= 4) -+ if (avr32_return_in_msb (type)) -+ /* Aggregates of size less than a word which does align the data in the -+ MSB must use SImode for r12. */ -+ return gen_rtx_REG (SImode, RET_REGISTER); -+ else -+ return gen_rtx_REG (TYPE_MODE (type), RET_REGISTER); -+ else if (int_size_in_bytes (type) <= 8) -+ return gen_rtx_REG (TYPE_MODE (type), INTERNAL_REGNUM (11)); -+ -+ return NULL_RTX; -+ } -+ -+/* -+ Return a rtx used for the return value from a library function call. -+ */ -+rtx -+avr32_libcall_value (enum machine_mode mode) -+ { -+ -+ if (GET_MODE_SIZE (mode) <= 4) -+ return gen_rtx_REG (mode, RET_REGISTER); -+ else if (GET_MODE_SIZE (mode) <= 8) -+ return gen_rtx_REG (mode, INTERNAL_REGNUM (11)); -+ else -+ return NULL_RTX; -+ } -+ -+/* Return TRUE if X references a SYMBOL_REF. */ -+int -+symbol_mentioned_p (rtx x) -+ { -+ const char *fmt; -+ int i; -+ -+ if (GET_CODE (x) == SYMBOL_REF) -+ return 1; -+ -+ fmt = GET_RTX_FORMAT (GET_CODE (x)); -+ -+ for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--) -+ { -+ if (fmt[i] == 'E') -+ { -+ int j; -+ -+ for (j = XVECLEN (x, i) - 1; j >= 0; j--) -+ if (symbol_mentioned_p (XVECEXP (x, i, j))) -+ return 1; -+ } -+ else if (fmt[i] == 'e' && symbol_mentioned_p (XEXP (x, i))) -+ return 1; -+ } -+ -+ return 0; -+ } -+ -+/* Return TRUE if X references a LABEL_REF. */ -+int -+label_mentioned_p (rtx x) -+ { -+ const char *fmt; -+ int i; -+ -+ if (GET_CODE (x) == LABEL_REF) -+ return 1; -+ -+ fmt = GET_RTX_FORMAT (GET_CODE (x)); -+ for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--) -+ { -+ if (fmt[i] == 'E') -+ { -+ int j; -+ -+ for (j = XVECLEN (x, i) - 1; j >= 0; j--) -+ if (label_mentioned_p (XVECEXP (x, i, j))) -+ return 1; -+ } -+ else if (fmt[i] == 'e' && label_mentioned_p (XEXP (x, i))) -+ return 1; -+ } -+ -+ return 0; -+ } -+ -+ -+int -+avr32_legitimate_pic_operand_p (rtx x) -+ { -+ -+ /* We can't have const, this must be broken down to a symbol. */ -+ if (GET_CODE (x) == CONST) -+ return FALSE; -+ -+ /* Can't access symbols or labels via the constant pool either */ -+ if ((GET_CODE (x) == SYMBOL_REF -+ && CONSTANT_POOL_ADDRESS_P (x) -+ && (symbol_mentioned_p (get_pool_constant (x)) -+ || label_mentioned_p (get_pool_constant (x))))) -+ return FALSE; -+ -+ return TRUE; -+ } -+ -+ -+rtx -+legitimize_pic_address (rtx orig, enum machine_mode mode ATTRIBUTE_UNUSED, -+ rtx reg) -+ { -+ -+ if (GET_CODE (orig) == SYMBOL_REF || GET_CODE (orig) == LABEL_REF) -+ { -+ int subregs = 0; -+ -+ if (reg == 0) -+ { -+ if (no_new_pseudos) -+ abort (); -+ else -+ reg = gen_reg_rtx (Pmode); -+ -+ subregs = 1; -+ } -+ -+ emit_move_insn (reg, orig); -+ -+ /* Only set current function as using pic offset table if flag_pic is -+ set. This is because this function is also used if -+ TARGET_HAS_ASM_ADDR_PSEUDOS is set. */ -+ if (flag_pic) -+ current_function_uses_pic_offset_table = 1; -+ -+ /* Put a REG_EQUAL note on this insn, so that it can be optimized by -+ loop. */ -+ return reg; -+ } -+ else if (GET_CODE (orig) == CONST) -+ { -+ rtx base, offset; -+ -+ if (flag_pic -+ && GET_CODE (XEXP (orig, 0)) == PLUS -+ && XEXP (XEXP (orig, 0), 0) == pic_offset_table_rtx) -+ return orig; -+ -+ if (reg == 0) -+ { -+ if (no_new_pseudos) -+ abort (); -+ else -+ reg = gen_reg_rtx (Pmode); -+ } -+ -+ if (GET_CODE (XEXP (orig, 0)) == PLUS) -+ { -+ base = -+ legitimize_pic_address (XEXP (XEXP (orig, 0), 0), Pmode, reg); -+ offset = -+ legitimize_pic_address (XEXP (XEXP (orig, 0), 1), Pmode, -+ base == reg ? 0 : reg); -+ } -+ else -+ abort (); -+ -+ if (GET_CODE (offset) == CONST_INT) -+ { -+ /* The base register doesn't really matter, we only want to test -+ the index for the appropriate mode. */ -+ if (!avr32_const_ok_for_constraint_p (INTVAL (offset), 'I', "Is21")) -+ { -+ if (!no_new_pseudos) -+ offset = force_reg (Pmode, offset); -+ else -+ abort (); -+ } -+ -+ if (GET_CODE (offset) == CONST_INT) -+ return plus_constant (base, INTVAL (offset)); -+ } -+ -+ return gen_rtx_PLUS (Pmode, base, offset); -+ } -+ -+ return orig; -+ } -+ -+/* Generate code to load the PIC register. */ -+void -+avr32_load_pic_register (void) -+ { -+ rtx l1, pic_tmp; -+ rtx global_offset_table; -+ -+ if ((current_function_uses_pic_offset_table == 0) || TARGET_NO_INIT_GOT) -+ return; -+ -+ if (!flag_pic) -+ abort (); -+ -+ l1 = gen_label_rtx (); -+ -+ global_offset_table = gen_rtx_SYMBOL_REF (Pmode, "_GLOBAL_OFFSET_TABLE_"); -+ pic_tmp = -+ gen_rtx_CONST (Pmode, -+ gen_rtx_MINUS (SImode, gen_rtx_LABEL_REF (Pmode, l1), -+ global_offset_table)); -+ emit_insn (gen_pic_load_addr -+ (pic_offset_table_rtx, force_const_mem (SImode, pic_tmp))); -+ emit_insn (gen_pic_compute_got_from_pc (pic_offset_table_rtx, l1)); -+ -+ /* Need to emit this whether or not we obey regdecls, since setjmp/longjmp -+ can cause life info to screw up. */ -+ emit_insn (gen_rtx_USE (VOIDmode, pic_offset_table_rtx)); -+ } -+ -+ -+ -+/* This hook should return true if values of type type are returned at the most -+ significant end of a register (in other words, if they are padded at the -+ least significant end). You can assume that type is returned in a register; -+ the caller is required to check this. Note that the register provided by -+ FUNCTION_VALUE must be able to hold the complete return value. For example, -+ if a 1-, 2- or 3-byte structure is returned at the most significant end of a -+ 4-byte register, FUNCTION_VALUE should provide an SImode rtx. */ -+bool -+avr32_return_in_msb (tree type ATTRIBUTE_UNUSED) -+ { -+ /* if ( AGGREGATE_TYPE_P (type) ) if ((int_size_in_bytes(type) == 1) || -+ ((int_size_in_bytes(type) == 2) && TYPE_ALIGN_UNIT(type) >= 2)) return -+ false; else return true; */ -+ -+ return false; -+ } -+ -+ -+/* -+ Returns one if a certain function value is going to be returned in memory -+ and zero if it is going to be returned in a register. -+ -+ BLKmode and all other modes that is larger than 64 bits are returned in -+ memory. -+ */ -+bool -+avr32_return_in_memory (tree type, tree fntype ATTRIBUTE_UNUSED) -+ { -+ if (TYPE_MODE (type) == VOIDmode) -+ return false; -+ -+ if (int_size_in_bytes (type) > (2 * UNITS_PER_WORD) -+ || int_size_in_bytes (type) == -1) -+ { -+ return true; -+ } -+ -+ /* If we have an aggregate then use the same mechanism as when checking if -+ it should be passed on the stack. */ -+ if (type -+ && AGGREGATE_TYPE_P (type) -+ && (*targetm.calls.must_pass_in_stack) (TYPE_MODE (type), type)) -+ return true; -+ -+ return false; -+ } -+ -+ -+/* Output the constant part of the trampoline. -+ lddpc r0, pc[0x8:e] ; load static chain register -+ lddpc pc, pc[0x8:e] ; jump to subrutine -+ .long 0 ; Address to static chain, -+ ; filled in by avr32_initialize_trampoline() -+ .long 0 ; Address to subrutine, -+ ; filled in by avr32_initialize_trampoline() -+ */ -+void -+avr32_trampoline_template (FILE * file) -+ { -+ fprintf (file, "\tlddpc r0, pc[8]\n"); -+ fprintf (file, "\tlddpc pc, pc[8]\n"); -+ /* make room for the address of the static chain. */ -+ fprintf (file, "\t.long\t0\n"); -+ /* make room for the address to the subrutine. */ -+ fprintf (file, "\t.long\t0\n"); -+ } -+ -+ -+/* -+ Initialize the variable parts of a trampoline. -+ */ -+void -+avr32_initialize_trampoline (rtx addr, rtx fnaddr, rtx static_chain) -+ { -+ /* Store the address to the static chain. */ -+ emit_move_insn (gen_rtx_MEM -+ (SImode, plus_constant (addr, TRAMPOLINE_SIZE - 4)), -+ static_chain); -+ -+ /* Store the address to the function. */ -+ emit_move_insn (gen_rtx_MEM (SImode, plus_constant (addr, TRAMPOLINE_SIZE)), -+ fnaddr); -+ -+ emit_insn (gen_cache (gen_rtx_REG (SImode, 13), -+ gen_rtx_CONST_INT (SImode, -+ AVR32_CACHE_INVALIDATE_ICACHE))); -+ } -+ -+/* Return nonzero if X is valid as an addressing register. */ -+int -+avr32_address_register_rtx_p (rtx x, int strict_p) -+ { -+ int regno; -+ -+ if (!register_operand(x, GET_MODE(x))) -+ return 0; -+ -+ /* If strict we require the register to be a hard register. */ -+ if (strict_p -+ && !REG_P(x)) -+ return 0; -+ -+ regno = REGNO (x); -+ -+ if (strict_p) -+ return REGNO_OK_FOR_BASE_P (regno); -+ -+ return (regno <= LAST_REGNUM || regno >= FIRST_PSEUDO_REGISTER); -+ } -+ -+/* Return nonzero if INDEX is valid for an address index operand. */ -+int -+avr32_legitimate_index_p (enum machine_mode mode, rtx index, int strict_p) -+ { -+ enum rtx_code code = GET_CODE (index); -+ -+ if (GET_MODE_SIZE (mode) > 8) -+ return 0; -+ -+ /* Standard coprocessor addressing modes. */ -+ if (code == CONST_INT) -+ { -+ if (TARGET_HARD_FLOAT && GET_MODE_CLASS (mode) == MODE_FLOAT) -+ /* Coprocessor mem insns has a smaller reach than ordinary mem insns */ -+ return CONST_OK_FOR_CONSTRAINT_P (INTVAL (index), 'K', "Ku14"); -+ else -+ return CONST_OK_FOR_CONSTRAINT_P (INTVAL (index), 'K', "Ks16"); -+ } -+ -+ if (avr32_address_register_rtx_p (index, strict_p)) -+ return 1; -+ -+ if (code == MULT) -+ { -+ rtx xiop0 = XEXP (index, 0); -+ rtx xiop1 = XEXP (index, 1); -+ return ((avr32_address_register_rtx_p (xiop0, strict_p) -+ && power_of_two_operand (xiop1, SImode) -+ && (INTVAL (xiop1) <= 8)) -+ || (avr32_address_register_rtx_p (xiop1, strict_p) -+ && power_of_two_operand (xiop0, SImode) -+ && (INTVAL (xiop0) <= 8))); -+ } -+ else if (code == ASHIFT) -+ { -+ rtx op = XEXP (index, 1); -+ -+ return (avr32_address_register_rtx_p (XEXP (index, 0), strict_p) -+ && GET_CODE (op) == CONST_INT -+ && INTVAL (op) > 0 && INTVAL (op) <= 3); -+ } -+ -+ return 0; -+ } -+ -+/* -+ Used in the GO_IF_LEGITIMATE_ADDRESS macro. Returns a nonzero value if -+ the RTX x is a legitimate memory address. -+ -+ Returns NO_REGS if the address is not legatime, GENERAL_REGS or ALL_REGS -+ if it is. -+ */ -+ -+/* Forward declaration*/ -+int is_minipool_label (rtx label); -+ -+int -+avr32_legitimate_address (enum machine_mode mode, rtx x, int strict) -+ { -+ -+ switch (GET_CODE (x)) -+ { -+ case REG: -+ return avr32_address_register_rtx_p (x, strict); -+ case CONST: -+ { -+ rtx label = avr32_find_symbol (x); -+ if (label -+ && -+ ((CONSTANT_POOL_ADDRESS_P (label) -+ && !(flag_pic -+ && (symbol_mentioned_p (get_pool_constant (label)) -+ || label_mentioned_p (get_pool_constant (label))))) -+ /* TODO! Can this ever happen??? */ -+ || ((GET_CODE (label) == LABEL_REF) -+ && GET_CODE (XEXP (label, 0)) == CODE_LABEL -+ && is_minipool_label (XEXP (label, 0))))) -+ { -+ return TRUE; -+ } -+ } -+ break; -+ case LABEL_REF: -+ if (GET_CODE (XEXP (x, 0)) == CODE_LABEL -+ && is_minipool_label (XEXP (x, 0))) -+ { -+ return TRUE; -+ } -+ break; -+ case SYMBOL_REF: -+ { -+ if (CONSTANT_POOL_ADDRESS_P (x) -+ && !(flag_pic -+ && (symbol_mentioned_p (get_pool_constant (x)) -+ || label_mentioned_p (get_pool_constant (x))))) -+ return TRUE; -+ /* -+ A symbol_ref is only legal if it is a function. If all of them are -+ legal, a pseudo reg that is a constant will be replaced by a -+ symbol_ref and make illegale code. SYMBOL_REF_FLAG is set by -+ ENCODE_SECTION_INFO. */ -+ else if (SYMBOL_REF_RCALL_FUNCTION_P (x)) -+ return TRUE; -+ break; -+ } -+ case PRE_DEC: /* (pre_dec (...)) */ -+ case POST_INC: /* (post_inc (...)) */ -+ return avr32_address_register_rtx_p (XEXP (x, 0), strict); -+ case PLUS: /* (plus (...) (...)) */ -+ { -+ rtx xop0 = XEXP (x, 0); -+ rtx xop1 = XEXP (x, 1); -+ -+ return ((avr32_address_register_rtx_p (xop0, strict) -+ && avr32_legitimate_index_p (mode, xop1, strict)) -+ || (avr32_address_register_rtx_p (xop1, strict) -+ && avr32_legitimate_index_p (mode, xop0, strict))); -+ } -+ default: -+ break; -+ } -+ -+ return FALSE; -+ } -+ -+ -+int -+avr32_const_double_immediate (rtx value) -+ { -+ HOST_WIDE_INT hi, lo; -+ -+ if (GET_CODE (value) != CONST_DOUBLE) -+ return FALSE; -+ -+ if (SCALAR_FLOAT_MODE_P (GET_MODE (value))) -+ { -+ HOST_WIDE_INT target_float[2]; -+ hi = lo = 0; -+ real_to_target (target_float, CONST_DOUBLE_REAL_VALUE (value), -+ GET_MODE (value)); -+ lo = target_float[0]; -+ hi = target_float[1]; -+ } -+ else -+ { -+ hi = CONST_DOUBLE_HIGH (value); -+ lo = CONST_DOUBLE_LOW (value); -+ } -+ -+ if (avr32_const_ok_for_constraint_p (lo, 'K', "Ks21") -+ && (GET_MODE (value) == SFmode -+ || avr32_const_ok_for_constraint_p (hi, 'K', "Ks21"))) -+ { -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+ -+int -+avr32_legitimate_constant_p (rtx x) -+ { -+ switch (GET_CODE (x)) -+ { -+ case CONST_INT: -+ /* Check if we should put large immediate into constant pool -+ or load them directly with mov/orh.*/ -+ if (!avr32_imm_in_const_pool) -+ return 1; -+ -+ return avr32_const_ok_for_constraint_p (INTVAL (x), 'K', "Ks21"); -+ case CONST_DOUBLE: -+ /* Check if we should put large immediate into constant pool -+ or load them directly with mov/orh.*/ -+ if (!avr32_imm_in_const_pool) -+ return 1; -+ -+ if (GET_MODE (x) == SFmode -+ || GET_MODE (x) == DFmode || GET_MODE (x) == DImode) -+ return avr32_const_double_immediate (x); -+ else -+ return 0; -+ case LABEL_REF: -+ return flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS; -+ case SYMBOL_REF: -+ return flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS; -+ case CONST: -+ case HIGH: -+ case CONST_VECTOR: -+ return 0; -+ default: -+ printf ("%s():\n", __FUNCTION__); -+ debug_rtx (x); -+ return 1; -+ } -+ } -+ -+ -+/* Strip any special encoding from labels */ -+const char * -+avr32_strip_name_encoding (const char *name) -+{ -+ const char *stripped = name; -+ -+ while (1) -+ { -+ switch (stripped[0]) -+ { -+ case '#': -+ stripped = strchr (name + 1, '#') + 1; -+ break; -+ case '*': -+ stripped = &stripped[1]; -+ break; -+ default: -+ return stripped; -+ } -+ } -+} -+ -+ -+ -+/* Do anything needed before RTL is emitted for each function. */ -+static struct machine_function * -+avr32_init_machine_status (void) -+{ -+ struct machine_function *machine; -+ machine = -+ (machine_function *) ggc_alloc_cleared (sizeof (machine_function)); -+ -+#if AVR32_FT_UNKNOWN != 0 -+ machine->func_type = AVR32_FT_UNKNOWN; -+#endif -+ -+ machine->minipool_label_head = 0; -+ machine->minipool_label_tail = 0; -+ return machine; -+} -+ -+void -+avr32_init_expanders (void) -+ { -+ /* Arrange to initialize and mark the machine per-function status. */ -+ init_machine_status = avr32_init_machine_status; -+ } -+ -+ -+/* Return an RTX indicating where the return address to the -+ calling function can be found. */ -+ -+rtx -+avr32_return_addr (int count, rtx frame ATTRIBUTE_UNUSED) -+ { -+ if (count != 0) -+ return NULL_RTX; -+ -+ return get_hard_reg_initial_val (Pmode, LR_REGNUM); -+ } -+ -+ -+void -+avr32_encode_section_info (tree decl, rtx rtl, int first) -+ { -+ -+ if (first && DECL_P (decl)) -+ { -+ /* Set SYMBOL_REG_FLAG for local functions */ -+ if (!TREE_PUBLIC (decl) && TREE_CODE (decl) == FUNCTION_DECL) -+ { -+ if ((*targetm.binds_local_p) (decl)) -+ { -+ SYMBOL_REF_FLAG (XEXP (rtl, 0)) = 1; -+ } -+ } -+ } -+ } -+ -+ -+void -+avr32_asm_output_ascii (FILE * stream, char *ptr, int len) -+ { -+ int i, i_new = 0; -+ char *new_ptr = xmalloc (4 * len); -+ if (new_ptr == NULL) -+ internal_error ("Out of memory."); -+ -+ for (i = 0; i < len; i++) -+ { -+ if (ptr[i] == '\n') -+ { -+ new_ptr[i_new++] = '\\'; -+ new_ptr[i_new++] = '0'; -+ new_ptr[i_new++] = '1'; -+ new_ptr[i_new++] = '2'; -+ } -+ else if (ptr[i] == '\"') -+ { -+ new_ptr[i_new++] = '\\'; -+ new_ptr[i_new++] = '\"'; -+ } -+ else if (ptr[i] == '\\') -+ { -+ new_ptr[i_new++] = '\\'; -+ new_ptr[i_new++] = '\\'; -+ } -+ else if (ptr[i] == '\0' && i + 1 < len) -+ { -+ new_ptr[i_new++] = '\\'; -+ new_ptr[i_new++] = '0'; -+ } -+ else -+ { -+ new_ptr[i_new++] = ptr[i]; -+ } -+ } -+ -+ /* Terminate new_ptr. */ -+ new_ptr[i_new] = '\0'; -+ fprintf (stream, "\t.ascii\t\"%s\"\n", new_ptr); -+ free (new_ptr); -+ } -+ -+ -+void -+avr32_asm_output_label (FILE * stream, const char *name) -+ { -+ name = avr32_strip_name_encoding (name); -+ -+ /* Print the label. */ -+ assemble_name (stream, name); -+ fprintf (stream, ":\n"); -+ } -+ -+ -+ -+void -+avr32_asm_weaken_label (FILE * stream, const char *name) -+ { -+ fprintf (stream, "\t.weak "); -+ assemble_name (stream, name); -+ fprintf (stream, "\n"); -+ } -+ -+/* -+ Checks if a labelref is equal to a reserved word in the assembler. If it is, -+ insert a '_' before the label name. -+ */ -+void -+avr32_asm_output_labelref (FILE * stream, const char *name) -+ { -+ int verbatim = FALSE; -+ const char *stripped = name; -+ int strip_finished = FALSE; -+ -+ while (!strip_finished) -+ { -+ switch (stripped[0]) -+ { -+ case '#': -+ stripped = strchr (name + 1, '#') + 1; -+ break; -+ case '*': -+ stripped = &stripped[1]; -+ verbatim = TRUE; -+ break; -+ default: -+ strip_finished = TRUE; -+ break; -+ } -+ } -+ -+ if (verbatim) -+ fputs (stripped, stream); -+ else -+ asm_fprintf (stream, "%U%s", stripped); -+ } -+ -+ -+ -+/* -+ Check if the comparison in compare_exp is redundant -+ for the condition given in next_cond given that the -+ needed flags are already set by an earlier instruction. -+ Uses cc_prev_status to check this. -+ -+ Returns NULL_RTX if the compare is not redundant -+ or the new condition to use in the conditional -+ instruction if the compare is redundant. -+ */ -+static rtx -+is_compare_redundant (rtx compare_exp, rtx next_cond) -+ { -+ int z_flag_valid = FALSE; -+ int n_flag_valid = FALSE; -+ rtx new_cond; -+ -+ if (GET_CODE (compare_exp) != COMPARE) -+ return NULL_RTX; -+ -+ -+ if (rtx_equal_p (cc_prev_status.mdep.value, compare_exp)) -+ { -+ /* cc0 already contains the correct comparison -> delete cmp insn */ -+ return next_cond; -+ } -+ -+ if (GET_MODE (compare_exp) != SImode) -+ return NULL_RTX; -+ -+ switch (cc_prev_status.mdep.flags) -+ { -+ case CC_SET_VNCZ: -+ case CC_SET_NCZ: -+ n_flag_valid = TRUE; -+ case CC_SET_CZ: -+ case CC_SET_Z: -+ z_flag_valid = TRUE; -+ } -+ -+ if (cc_prev_status.mdep.value -+ && REG_P (XEXP (compare_exp, 0)) -+ && REGNO (XEXP (compare_exp, 0)) == REGNO (cc_prev_status.mdep.value) -+ && GET_CODE (XEXP (compare_exp, 1)) == CONST_INT -+ && next_cond != NULL_RTX) -+ { -+ if (INTVAL (XEXP (compare_exp, 1)) == 0 -+ && z_flag_valid -+ && (GET_CODE (next_cond) == EQ || GET_CODE (next_cond) == NE)) -+ /* We can skip comparison Z flag is already reflecting ops[0] */ -+ return next_cond; -+ else if (n_flag_valid -+ && ((INTVAL (XEXP (compare_exp, 1)) == 0 -+ && (GET_CODE (next_cond) == GE -+ || GET_CODE (next_cond) == LT)) -+ || (INTVAL (XEXP (compare_exp, 1)) == -1 -+ && (GET_CODE (next_cond) == GT -+ || GET_CODE (next_cond) == LE)))) -+ { -+ /* We can skip comparison N flag is already reflecting ops[0], -+ which means that we can use the mi/pl conditions to check if -+ ops[0] is GE or LT 0. */ -+ if ((GET_CODE (next_cond) == GE) || (GET_CODE (next_cond) == GT)) -+ new_cond = -+ gen_rtx_UNSPEC (GET_MODE (next_cond), gen_rtvec (2, cc0_rtx, const0_rtx), -+ UNSPEC_COND_PL); -+ else -+ new_cond = -+ gen_rtx_UNSPEC (GET_MODE (next_cond), gen_rtvec (2, cc0_rtx, const0_rtx), -+ UNSPEC_COND_MI); -+ return new_cond; -+ } -+ } -+ return NULL_RTX; -+ } -+ -+/* Updates cc_status. */ -+void -+avr32_notice_update_cc (rtx exp, rtx insn) -+ { -+ switch (get_attr_cc (insn)) -+ { -+ case CC_CALL_SET: -+ CC_STATUS_INIT; -+ FPCC_STATUS_INIT; -+ /* Check if the function call returns a value in r12 */ -+ if (REG_P (recog_data.operand[0]) -+ && REGNO (recog_data.operand[0]) == RETVAL_REGNUM) -+ { -+ cc_status.flags = 0; -+ cc_status.mdep.value = -+ gen_rtx_COMPARE (SImode, recog_data.operand[0], const0_rtx); -+ cc_status.mdep.flags = CC_SET_VNCZ; -+ -+ } -+ break; -+ case CC_COMPARE: -+ /* Check that compare will not be optimized away if so nothing should -+ be done */ -+ if (is_compare_redundant (SET_SRC (exp), get_next_insn_cond (insn)) == -+ NULL_RTX) -+ { -+ -+ /* Reset the nonstandard flag */ -+ CC_STATUS_INIT; -+ cc_status.flags = 0; -+ cc_status.mdep.value = SET_SRC (exp); -+ cc_status.mdep.flags = CC_SET_VNCZ; -+ } -+ break; -+ case CC_CMP_COND_INSN: -+ { -+ /* Conditional insn that emit the compare itself. */ -+ rtx cmp = gen_rtx_COMPARE (GET_MODE (recog_data.operand[4]), -+ recog_data.operand[4], -+ recog_data.operand[5]); -+ -+ if (is_compare_redundant (cmp, recog_data.operand[1]) == NULL_RTX) -+ { -+ -+ /* Reset the nonstandard flag */ -+ CC_STATUS_INIT; -+ cc_status.flags = 0; -+ cc_status.mdep.value = cmp; -+ cc_status.mdep.flags = CC_SET_VNCZ; -+ } -+ } -+ break; -+ case CC_FPCOMPARE: -+ /* Check that floating-point compare will not be optimized away if so -+ nothing should be done */ -+ if (!rtx_equal_p (cc_prev_status.mdep.fpvalue, SET_SRC (exp))) -+ { -+ /* cc0 already contains the correct comparison -> delete cmp insn */ -+ /* Reset the nonstandard flag */ -+ cc_status.mdep.fpvalue = SET_SRC (exp); -+ cc_status.mdep.fpflags = CC_SET_CZ; -+ } -+ break; -+ case CC_FROM_FPCC: -+ /* Flags are updated with flags from Floating-point coprocessor, set -+ CC_NOT_SIGNED flag since the flags are set so that unsigned -+ condidion codes can be used directly. */ -+ CC_STATUS_INIT; -+ cc_status.flags = CC_NOT_SIGNED; -+ cc_status.mdep.value = cc_status.mdep.fpvalue; -+ cc_status.mdep.flags = cc_status.mdep.fpflags; -+ break; -+ case CC_BLD: -+ /* Bit load is kind of like an inverted testsi, because the Z flag is -+ inverted */ -+ CC_STATUS_INIT; -+ cc_status.flags = CC_INVERTED; -+ cc_status.mdep.value = SET_SRC (exp); -+ cc_status.mdep.flags = CC_SET_Z; -+ break; -+ case CC_NONE: -+ /* Insn does not affect CC at all. Check if the instruction updates -+ some of the register currently reflected in cc0 */ -+ -+ if ((GET_CODE (exp) == SET) -+ && (cc_status.value1 || cc_status.value2 || cc_status.mdep.value) -+ && (reg_mentioned_p (SET_DEST (exp), cc_status.value1) -+ || reg_mentioned_p (SET_DEST (exp), cc_status.value2) -+ || reg_mentioned_p (SET_DEST (exp), cc_status.mdep.value))) -+ { -+ CC_STATUS_INIT; -+ } -+ -+ /* If this is a parallel we must step through each of the parallel -+ expressions */ -+ if (GET_CODE (exp) == PARALLEL) -+ { -+ int i; -+ for (i = 0; i < XVECLEN (exp, 0); ++i) -+ { -+ rtx vec_exp = XVECEXP (exp, 0, i); -+ if ((GET_CODE (vec_exp) == SET) -+ && (cc_status.value1 || cc_status.value2 -+ || cc_status.mdep.value) -+ && (reg_mentioned_p (SET_DEST (vec_exp), cc_status.value1) -+ || reg_mentioned_p (SET_DEST (vec_exp), -+ cc_status.value2) -+ || reg_mentioned_p (SET_DEST (vec_exp), -+ cc_status.mdep.value))) -+ { -+ CC_STATUS_INIT; -+ } -+ } -+ } -+ -+ /* Check if we have memory opartions with post_inc or pre_dec on the -+ register currently reflected in cc0 */ -+ if (GET_CODE (exp) == SET -+ && GET_CODE (SET_SRC (exp)) == MEM -+ && (GET_CODE (XEXP (SET_SRC (exp), 0)) == POST_INC -+ || GET_CODE (XEXP (SET_SRC (exp), 0)) == PRE_DEC) -+ && -+ (reg_mentioned_p -+ (XEXP (XEXP (SET_SRC (exp), 0), 0), cc_status.value1) -+ || reg_mentioned_p (XEXP (XEXP (SET_SRC (exp), 0), 0), -+ cc_status.value2) -+ || reg_mentioned_p (XEXP (XEXP (SET_SRC (exp), 0), 0), -+ cc_status.mdep.value))) -+ CC_STATUS_INIT; -+ -+ if (GET_CODE (exp) == SET -+ && GET_CODE (SET_DEST (exp)) == MEM -+ && (GET_CODE (XEXP (SET_DEST (exp), 0)) == POST_INC -+ || GET_CODE (XEXP (SET_DEST (exp), 0)) == PRE_DEC) -+ && -+ (reg_mentioned_p -+ (XEXP (XEXP (SET_DEST (exp), 0), 0), cc_status.value1) -+ || reg_mentioned_p (XEXP (XEXP (SET_DEST (exp), 0), 0), -+ cc_status.value2) -+ || reg_mentioned_p (XEXP (XEXP (SET_DEST (exp), 0), 0), -+ cc_status.mdep.value))) -+ CC_STATUS_INIT; -+ break; -+ -+ case CC_SET_VNCZ: -+ CC_STATUS_INIT; -+ cc_status.mdep.value = recog_data.operand[0]; -+ cc_status.mdep.flags = CC_SET_VNCZ; -+ break; -+ -+ case CC_SET_NCZ: -+ CC_STATUS_INIT; -+ cc_status.mdep.value = recog_data.operand[0]; -+ cc_status.mdep.flags = CC_SET_NCZ; -+ break; -+ -+ case CC_SET_CZ: -+ CC_STATUS_INIT; -+ cc_status.mdep.value = recog_data.operand[0]; -+ cc_status.mdep.flags = CC_SET_CZ; -+ break; -+ -+ case CC_SET_Z: -+ CC_STATUS_INIT; -+ cc_status.mdep.value = recog_data.operand[0]; -+ cc_status.mdep.flags = CC_SET_Z; -+ break; -+ -+ case CC_CLOBBER: -+ CC_STATUS_INIT; -+ break; -+ -+ default: -+ CC_STATUS_INIT; -+ } -+ } -+ -+ -+/* -+ Outputs to stdio stream stream the assembler syntax for an instruction -+ operand x. x is an RTL expression. -+ */ -+void -+avr32_print_operand (FILE * stream, rtx x, int code) -+ { -+ int error = 0; -+ -+ switch (GET_CODE (x)) -+ { -+ case UNSPEC: -+ switch (XINT (x, 1)) -+ { -+ case UNSPEC_COND_PL: -+ if (code == 'i') -+ fputs ("mi", stream); -+ else -+ fputs ("pl", stream); -+ break; -+ case UNSPEC_COND_MI: -+ if (code == 'i') -+ fputs ("pl", stream); -+ else -+ fputs ("mi", stream); -+ break; -+ default: -+ error = 1; -+ } -+ break; -+ case EQ: -+ if (code == 'i') -+ fputs ("ne", stream); -+ else -+ fputs ("eq", stream); -+ break; -+ case NE: -+ if (code == 'i') -+ fputs ("eq", stream); -+ else -+ fputs ("ne", stream); -+ break; -+ case GT: -+ if (code == 'i') -+ fputs ("le", stream); -+ else -+ fputs ("gt", stream); -+ break; -+ case GTU: -+ if (code == 'i') -+ fputs ("ls", stream); -+ else -+ fputs ("hi", stream); -+ break; -+ case LT: -+ if (code == 'i') -+ fputs ("ge", stream); -+ else -+ fputs ("lt", stream); -+ break; -+ case LTU: -+ if (code == 'i') -+ fputs ("hs", stream); -+ else -+ fputs ("lo", stream); -+ break; -+ case GE: -+ if (code == 'i') -+ fputs ("lt", stream); -+ else -+ fputs ("ge", stream); -+ break; -+ case GEU: -+ if (code == 'i') -+ fputs ("lo", stream); -+ else -+ fputs ("hs", stream); -+ break; -+ case LE: -+ if (code == 'i') -+ fputs ("gt", stream); -+ else -+ fputs ("le", stream); -+ break; -+ case LEU: -+ if (code == 'i') -+ fputs ("hi", stream); -+ else -+ fputs ("ls", stream); -+ break; -+ case CONST_INT: -+ { -+ HOST_WIDE_INT value = INTVAL (x); -+ -+ switch (code) -+ { -+ case 'm': -+ if ( HOST_BITS_PER_WIDE_INT > BITS_PER_WORD ) -+ { -+ /* A const_int can be used to represent DImode constants. */ -+ value >>= BITS_PER_WORD; -+ } -+ /* We might get a const_int immediate for setting a DI register, -+ we then must then return the correct sign extended DI. The most -+ significant word is just a sign extension. */ -+ else if (value < 0) -+ value = -1; -+ else -+ value = 0; -+ break; -+ case 'i': -+ value++; -+ break; -+ case 'p': -+ { -+ /* Set to bit position of first bit set in immediate */ -+ int i, bitpos = 32; -+ for (i = 0; i < 32; i++) -+ if (value & (1 << i)) -+ { -+ bitpos = i; -+ break; -+ } -+ value = bitpos; -+ } -+ break; -+ case 'r': -+ { -+ /* Reglist 8 */ -+ char op[50]; -+ op[0] = '\0'; -+ -+ if (value & 0x01) -+ sprintf (op, "r0-r3"); -+ if (value & 0x02) -+ strlen (op) ? sprintf (op, "%s, r4-r7", op) : sprintf (op, -+ "r4-r7"); -+ if (value & 0x04) -+ strlen (op) ? sprintf (op, "%s, r8-r9", op) : sprintf (op, -+ "r8-r9"); -+ if (value & 0x08) -+ strlen (op) ? sprintf (op, "%s, r10", op) : sprintf (op, -+ "r10"); -+ if (value & 0x10) -+ strlen (op) ? sprintf (op, "%s, r11", op) : sprintf (op, -+ "r11"); -+ if (value & 0x20) -+ strlen (op) ? sprintf (op, "%s, r12", op) : sprintf (op, -+ "r12"); -+ if (value & 0x40) -+ strlen (op) ? sprintf (op, "%s, lr", op) : sprintf (op, "lr"); -+ if (value & 0x80) -+ strlen (op) ? sprintf (op, "%s, pc", op) : sprintf (op, "pc"); -+ -+ fputs (op, stream); -+ return; -+ } -+ case 's': -+ { -+ /* Reglist 16 */ -+ char reglist16_string[100]; -+ int i; -+ reglist16_string[0] = '\0'; -+ -+ for (i = 0; i < 16; ++i) -+ { -+ if (value & (1 << i)) -+ { -+ strlen (reglist16_string) ? sprintf (reglist16_string, -+ "%s, %s", -+ reglist16_string, -+ reg_names -+ [INTERNAL_REGNUM -+ (i)]) : -+ sprintf (reglist16_string, "%s", -+ reg_names[INTERNAL_REGNUM (i)]); -+ } -+ } -+ fputs (reglist16_string, stream); -+ return; -+ } -+ case 'C': -+ { -+ /* RegListCP8 */ -+ char reglist_string[100]; -+ avr32_make_fp_reglist_w (value, (char *) reglist_string); -+ fputs (reglist_string, stream); -+ return; -+ } -+ case 'D': -+ { -+ /* RegListCPD8 */ -+ char reglist_string[100]; -+ avr32_make_fp_reglist_d (value, (char *) reglist_string); -+ fputs (reglist_string, stream); -+ return; -+ } -+ case 'h': -+ /* Print halfword part of word */ -+ fputs (value ? "b" : "t", stream); -+ return; -+ } -+ -+ /* Print Value */ -+ fprintf (stream, "%d", value); -+ break; -+ } -+ case CONST_DOUBLE: -+ { -+ HOST_WIDE_INT hi, lo; -+ if (SCALAR_FLOAT_MODE_P (GET_MODE (x))) -+ { -+ HOST_WIDE_INT target_float[2]; -+ hi = lo = 0; -+ real_to_target (target_float, CONST_DOUBLE_REAL_VALUE (x), -+ GET_MODE (x)); -+ /* For doubles the most significant part starts at index 0. */ -+ if (GET_MODE_SIZE (GET_MODE (x)) > UNITS_PER_WORD) -+ { -+ hi = target_float[0]; -+ lo = target_float[1]; -+ } -+ else -+ { -+ lo = target_float[0]; -+ } -+ } -+ else -+ { -+ hi = CONST_DOUBLE_HIGH (x); -+ lo = CONST_DOUBLE_LOW (x); -+ } -+ -+ if (code == 'm') -+ fprintf (stream, "%ld", hi); -+ else -+ fprintf (stream, "%ld", lo); -+ -+ break; -+ } -+ case CONST: -+ output_addr_const (stream, XEXP (XEXP (x, 0), 0)); -+ fprintf (stream, "+%ld", INTVAL (XEXP (XEXP (x, 0), 1))); -+ break; -+ case REG: -+ /* Swap register name if the register is DImode or DFmode. */ -+ if (GET_MODE (x) == DImode || GET_MODE (x) == DFmode) -+ { -+ /* Double register must have an even numbered address */ -+ gcc_assert (!(REGNO (x) % 2)); -+ if (code == 'm') -+ fputs (reg_names[true_regnum (x)], stream); -+ else -+ fputs (reg_names[true_regnum (x) + 1], stream); -+ } -+ else if (GET_MODE (x) == TImode) -+ { -+ switch (code) -+ { -+ case 'T': -+ fputs (reg_names[true_regnum (x)], stream); -+ break; -+ case 'U': -+ fputs (reg_names[true_regnum (x) + 1], stream); -+ break; -+ case 'L': -+ fputs (reg_names[true_regnum (x) + 2], stream); -+ break; -+ case 'B': -+ fputs (reg_names[true_regnum (x) + 3], stream); -+ break; -+ default: -+ fprintf (stream, "%s, %s, %s, %s", -+ reg_names[true_regnum (x) + 3], -+ reg_names[true_regnum (x) + 2], -+ reg_names[true_regnum (x) + 1], -+ reg_names[true_regnum (x)]); -+ break; -+ } -+ } -+ else -+ { -+ fputs (reg_names[true_regnum (x)], stream); -+ } -+ break; -+ case CODE_LABEL: -+ case LABEL_REF: -+ case SYMBOL_REF: -+ output_addr_const (stream, x); -+ break; -+ case MEM: -+ switch (GET_CODE (XEXP (x, 0))) -+ { -+ case LABEL_REF: -+ case SYMBOL_REF: -+ output_addr_const (stream, XEXP (x, 0)); -+ break; -+ case MEM: -+ switch (GET_CODE (XEXP (XEXP (x, 0), 0))) -+ { -+ case SYMBOL_REF: -+ output_addr_const (stream, XEXP (XEXP (x, 0), 0)); -+ break; -+ default: -+ error = 1; -+ break; -+ } -+ break; -+ case REG: -+ avr32_print_operand (stream, XEXP (x, 0), 0); -+ if (code != 'p') -+ fputs ("[0]", stream); -+ break; -+ case PRE_DEC: -+ fputs ("--", stream); -+ avr32_print_operand (stream, XEXP (XEXP (x, 0), 0), 0); -+ break; -+ case POST_INC: -+ avr32_print_operand (stream, XEXP (XEXP (x, 0), 0), 0); -+ fputs ("++", stream); -+ break; -+ case PLUS: -+ { -+ rtx op0 = XEXP (XEXP (x, 0), 0); -+ rtx op1 = XEXP (XEXP (x, 0), 1); -+ rtx base = NULL_RTX, offset = NULL_RTX; -+ -+ if (avr32_address_register_rtx_p (op0, 1)) -+ { -+ base = op0; -+ offset = op1; -+ } -+ else if (avr32_address_register_rtx_p (op1, 1)) -+ { -+ /* Operands are switched. */ -+ base = op1; -+ offset = op0; -+ } -+ -+ gcc_assert (base && offset -+ && avr32_address_register_rtx_p (base, 1) -+ && avr32_legitimate_index_p (GET_MODE (x), offset, -+ 1)); -+ -+ avr32_print_operand (stream, base, 0); -+ fputs ("[", stream); -+ avr32_print_operand (stream, offset, 0); -+ fputs ("]", stream); -+ break; -+ } -+ case CONST: -+ output_addr_const (stream, XEXP (XEXP (XEXP (x, 0), 0), 0)); -+ fprintf (stream, " + %ld", -+ INTVAL (XEXP (XEXP (XEXP (x, 0), 0), 1))); -+ break; -+ default: -+ error = 1; -+ } -+ break; -+ case MULT: -+ { -+ int value = INTVAL (XEXP (x, 1)); -+ -+ /* Convert immediate in multiplication into a shift immediate */ -+ switch (value) -+ { -+ case 2: -+ value = 1; -+ break; -+ case 4: -+ value = 2; -+ break; -+ case 8: -+ value = 3; -+ break; -+ default: -+ value = 0; -+ } -+ fprintf (stream, "%s << %i", reg_names[true_regnum (XEXP (x, 0))], -+ value); -+ break; -+ } -+ case ASHIFT: -+ if (GET_CODE (XEXP (x, 1)) == CONST_INT) -+ fprintf (stream, "%s << %i", reg_names[true_regnum (XEXP (x, 0))], -+ (int) INTVAL (XEXP (x, 1))); -+ else if (REG_P (XEXP (x, 1))) -+ fprintf (stream, "%s << %s", reg_names[true_regnum (XEXP (x, 0))], -+ reg_names[true_regnum (XEXP (x, 1))]); -+ else -+ { -+ error = 1; -+ } -+ break; -+ case LSHIFTRT: -+ if (GET_CODE (XEXP (x, 1)) == CONST_INT) -+ fprintf (stream, "%s >> %i", reg_names[true_regnum (XEXP (x, 0))], -+ (int) INTVAL (XEXP (x, 1))); -+ else if (REG_P (XEXP (x, 1))) -+ fprintf (stream, "%s >> %s", reg_names[true_regnum (XEXP (x, 0))], -+ reg_names[true_regnum (XEXP (x, 1))]); -+ else -+ { -+ error = 1; -+ } -+ fprintf (stream, ">>"); -+ break; -+ case PARALLEL: -+ { -+ /* Load store multiple */ -+ int i; -+ int count = XVECLEN (x, 0); -+ int reglist16 = 0; -+ char reglist16_string[100]; -+ -+ for (i = 0; i < count; ++i) -+ { -+ rtx vec_elm = XVECEXP (x, 0, i); -+ if (GET_MODE (vec_elm) != SET) -+ { -+ debug_rtx (vec_elm); -+ internal_error ("Unknown element in parallel expression!"); -+ } -+ if (GET_MODE (XEXP (vec_elm, 0)) == REG) -+ { -+ /* Load multiple */ -+ reglist16 |= 1 << ASM_REGNUM (REGNO (XEXP (vec_elm, 0))); -+ } -+ else -+ { -+ /* Store multiple */ -+ reglist16 |= 1 << ASM_REGNUM (REGNO (XEXP (vec_elm, 1))); -+ } -+ } -+ -+ avr32_make_reglist16 (reglist16, reglist16_string); -+ fputs (reglist16_string, stream); -+ -+ break; -+ } -+ -+ default: -+ error = 1; -+ } -+ -+ if (error) -+ { -+ debug_rtx (x); -+ internal_error ("Illegal expression for avr32_print_operand"); -+ } -+ } -+ -+rtx -+avr32_get_note_reg_equiv (rtx insn) -+ { -+ rtx note; -+ -+ note = find_reg_note (insn, REG_EQUIV, NULL_RTX); -+ -+ if (note != NULL_RTX) -+ return XEXP (note, 0); -+ else -+ return NULL_RTX; -+ } -+ -+/* -+ Outputs to stdio stream stream the assembler syntax for an instruction -+ operand that is a memory reference whose address is x. x is an RTL -+ expression. -+ -+ ToDo: fixme. -+ */ -+void -+avr32_print_operand_address (FILE * stream, rtx x) -+ { -+ fprintf (stream, "(%d) /* address */", REGNO (x)); -+ } -+ -+/* Return true if _GLOBAL_OFFSET_TABLE_ symbol is mentioned. */ -+bool -+avr32_got_mentioned_p (rtx addr) -+ { -+ if (GET_CODE (addr) == MEM) -+ addr = XEXP (addr, 0); -+ while (GET_CODE (addr) == CONST) -+ addr = XEXP (addr, 0); -+ if (GET_CODE (addr) == SYMBOL_REF) -+ { -+ return streq (XSTR (addr, 0), "_GLOBAL_OFFSET_TABLE_"); -+ } -+ if (GET_CODE (addr) == PLUS || GET_CODE (addr) == MINUS) -+ { -+ bool l1, l2; -+ -+ l1 = avr32_got_mentioned_p (XEXP (addr, 0)); -+ l2 = avr32_got_mentioned_p (XEXP (addr, 1)); -+ return l1 || l2; -+ } -+ return false; -+ } -+ -+ -+/* Find the symbol in an address expression. */ -+ -+rtx -+avr32_find_symbol (rtx addr) -+ { -+ if (GET_CODE (addr) == MEM) -+ addr = XEXP (addr, 0); -+ -+ while (GET_CODE (addr) == CONST) -+ addr = XEXP (addr, 0); -+ -+ if (GET_CODE (addr) == SYMBOL_REF || GET_CODE (addr) == LABEL_REF) -+ return addr; -+ if (GET_CODE (addr) == PLUS) -+ { -+ rtx l1, l2; -+ -+ l1 = avr32_find_symbol (XEXP (addr, 0)); -+ l2 = avr32_find_symbol (XEXP (addr, 1)); -+ if (l1 != NULL_RTX && l2 == NULL_RTX) -+ return l1; -+ else if (l1 == NULL_RTX && l2 != NULL_RTX) -+ return l2; -+ } -+ -+ return NULL_RTX; -+ } -+ -+ -+/* Routines for manipulation of the constant pool. */ -+ -+/* AVR32 instructions cannot load a large constant directly into a -+ register; they have to come from a pc relative load. The constant -+ must therefore be placed in the addressable range of the pc -+ relative load. Depending on the precise pc relative load -+ instruction the range is somewhere between 256 bytes and 4k. This -+ means that we often have to dump a constant inside a function, and -+ generate code to branch around it. -+ -+ It is important to minimize this, since the branches will slow -+ things down and make the code larger. -+ -+ Normally we can hide the table after an existing unconditional -+ branch so that there is no interruption of the flow, but in the -+ worst case the code looks like this: -+ -+ lddpc rn, L1 -+ ... -+ rjmp L2 -+ align -+ L1: .long value -+ L2: -+ ... -+ -+ lddpc rn, L3 -+ ... -+ rjmp L4 -+ align -+ L3: .long value -+ L4: -+ ... -+ -+ We fix this by performing a scan after scheduling, which notices -+ which instructions need to have their operands fetched from the -+ constant table and builds the table. -+ -+ The algorithm starts by building a table of all the constants that -+ need fixing up and all the natural barriers in the function (places -+ where a constant table can be dropped without breaking the flow). -+ For each fixup we note how far the pc-relative replacement will be -+ able to reach and the offset of the instruction into the function. -+ -+ Having built the table we then group the fixes together to form -+ tables that are as large as possible (subject to addressing -+ constraints) and emit each table of constants after the last -+ barrier that is within range of all the instructions in the group. -+ If a group does not contain a barrier, then we forcibly create one -+ by inserting a jump instruction into the flow. Once the table has -+ been inserted, the insns are then modified to reference the -+ relevant entry in the pool. -+ -+ Possible enhancements to the algorithm (not implemented) are: -+ -+ 1) For some processors and object formats, there may be benefit in -+ aligning the pools to the start of cache lines; this alignment -+ would need to be taken into account when calculating addressability -+ of a pool. */ -+ -+/* These typedefs are located at the start of this file, so that -+ they can be used in the prototypes there. This comment is to -+ remind readers of that fact so that the following structures -+ can be understood more easily. -+ -+ typedef struct minipool_node Mnode; -+ typedef struct minipool_fixup Mfix; */ -+ -+struct minipool_node -+{ -+ /* Doubly linked chain of entries. */ -+ Mnode *next; -+ Mnode *prev; -+ /* The maximum offset into the code that this entry can be placed. While -+ pushing fixes for forward references, all entries are sorted in order of -+ increasing max_address. */ -+ HOST_WIDE_INT max_address; -+ /* Similarly for an entry inserted for a backwards ref. */ -+ HOST_WIDE_INT min_address; -+ /* The number of fixes referencing this entry. This can become zero if we -+ "unpush" an entry. In this case we ignore the entry when we come to -+ emit the code. */ -+ int refcount; -+ /* The offset from the start of the minipool. */ -+ HOST_WIDE_INT offset; -+ /* The value in table. */ -+ rtx value; -+ /* The mode of value. */ -+ enum machine_mode mode; -+ /* The size of the value. */ -+ int fix_size; -+}; -+ -+struct minipool_fixup -+{ -+ Mfix *next; -+ rtx insn; -+ HOST_WIDE_INT address; -+ rtx *loc; -+ enum machine_mode mode; -+ int fix_size; -+ rtx value; -+ Mnode *minipool; -+ HOST_WIDE_INT forwards; -+ HOST_WIDE_INT backwards; -+}; -+ -+ -+/* Fixes less than a word need padding out to a word boundary. */ -+#define MINIPOOL_FIX_SIZE(mode, value) \ -+ (IS_FORCE_MINIPOOL(value) ? 0 : \ -+ (GET_MODE_SIZE ((mode)) >= 4 ? GET_MODE_SIZE ((mode)) : 4)) -+ -+#define IS_FORCE_MINIPOOL(x) \ -+ (GET_CODE(x) == UNSPEC && \ -+ XINT(x, 1) == UNSPEC_FORCE_MINIPOOL) -+ -+static Mnode *minipool_vector_head; -+static Mnode *minipool_vector_tail; -+ -+/* The linked list of all minipool fixes required for this function. */ -+Mfix *minipool_fix_head; -+Mfix *minipool_fix_tail; -+/* The fix entry for the current minipool, once it has been placed. */ -+Mfix *minipool_barrier; -+ -+/* Determines if INSN is the start of a jump table. Returns the end -+ of the TABLE or NULL_RTX. */ -+static rtx -+is_jump_table (rtx insn) -+ { -+ rtx table; -+ -+ if (GET_CODE (insn) == JUMP_INSN -+ && JUMP_LABEL (insn) != NULL -+ && ((table = next_real_insn (JUMP_LABEL (insn))) -+ == next_real_insn (insn)) -+ && table != NULL -+ && GET_CODE (table) == JUMP_INSN -+ && (GET_CODE (PATTERN (table)) == ADDR_VEC -+ || GET_CODE (PATTERN (table)) == ADDR_DIFF_VEC)) -+ return table; -+ -+ return NULL_RTX; -+ } -+ -+static HOST_WIDE_INT -+get_jump_table_size (rtx insn) -+ { -+ /* ADDR_VECs only take room if read-only data does into the text section. */ -+ if (JUMP_TABLES_IN_TEXT_SECTION -+#if !defined(READONLY_DATA_SECTION_ASM_OP) -+ || 1 -+#endif -+ ) -+ { -+ rtx body = PATTERN (insn); -+ int elt = GET_CODE (body) == ADDR_DIFF_VEC ? 1 : 0; -+ -+ return GET_MODE_SIZE (GET_MODE (body)) * XVECLEN (body, elt); -+ } -+ -+ return 0; -+ } -+ -+/* Move a minipool fix MP from its current location to before MAX_MP. -+ If MAX_MP is NULL, then MP doesn't need moving, but the addressing -+ constraints may need updating. */ -+static Mnode * -+move_minipool_fix_forward_ref (Mnode * mp, Mnode * max_mp, -+ HOST_WIDE_INT max_address) -+ { -+ /* This should never be true and the code below assumes these are -+ different. */ -+ if (mp == max_mp) -+ abort (); -+ -+ if (max_mp == NULL) -+ { -+ if (max_address < mp->max_address) -+ mp->max_address = max_address; -+ } -+ else -+ { -+ if (max_address > max_mp->max_address - mp->fix_size) -+ mp->max_address = max_mp->max_address - mp->fix_size; -+ else -+ mp->max_address = max_address; -+ -+ /* Unlink MP from its current position. Since max_mp is non-null, -+ mp->prev must be non-null. */ -+ mp->prev->next = mp->next; -+ if (mp->next != NULL) -+ mp->next->prev = mp->prev; -+ else -+ minipool_vector_tail = mp->prev; -+ -+ /* Re-insert it before MAX_MP. */ -+ mp->next = max_mp; -+ mp->prev = max_mp->prev; -+ max_mp->prev = mp; -+ -+ if (mp->prev != NULL) -+ mp->prev->next = mp; -+ else -+ minipool_vector_head = mp; -+ } -+ -+ /* Save the new entry. */ -+ max_mp = mp; -+ -+ /* Scan over the preceding entries and adjust their addresses as required. -+ */ -+ while (mp->prev != NULL -+ && mp->prev->max_address > mp->max_address - mp->prev->fix_size) -+ { -+ mp->prev->max_address = mp->max_address - mp->prev->fix_size; -+ mp = mp->prev; -+ } -+ -+ return max_mp; -+ } -+ -+/* Add a constant to the minipool for a forward reference. Returns the -+ node added or NULL if the constant will not fit in this pool. */ -+static Mnode * -+add_minipool_forward_ref (Mfix * fix) -+{ -+ /* If set, max_mp is the first pool_entry that has a lower constraint than -+ the one we are trying to add. */ -+ Mnode *max_mp = NULL; -+ HOST_WIDE_INT max_address = fix->address + fix->forwards; -+ Mnode *mp; -+ -+ /* If this fix's address is greater than the address of the first entry, -+ then we can't put the fix in this pool. We subtract the size of the -+ current fix to ensure that if the table is fully packed we still have -+ enough room to insert this value by suffling the other fixes forwards. */ -+ if (minipool_vector_head && -+ fix->address >= minipool_vector_head->max_address - fix->fix_size) -+ return NULL; -+ -+ /* Scan the pool to see if a constant with the same value has already been -+ added. While we are doing this, also note the location where we must -+ insert the constant if it doesn't already exist. */ -+ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) -+ { -+ if (GET_CODE (fix->value) == GET_CODE (mp->value) -+ && fix->mode == mp->mode -+ && (GET_CODE (fix->value) != CODE_LABEL -+ || (CODE_LABEL_NUMBER (fix->value) -+ == CODE_LABEL_NUMBER (mp->value))) -+ && rtx_equal_p (fix->value, mp->value)) -+ { -+ /* More than one fix references this entry. */ -+ mp->refcount++; -+ return move_minipool_fix_forward_ref (mp, max_mp, max_address); -+ } -+ -+ /* Note the insertion point if necessary. */ -+ if (max_mp == NULL && mp->max_address > max_address) -+ max_mp = mp; -+ -+ } -+ -+ /* The value is not currently in the minipool, so we need to create a new -+ entry for it. If MAX_MP is NULL, the entry will be put on the end of -+ the list since the placement is less constrained than any existing -+ entry. Otherwise, we insert the new fix before MAX_MP and, if -+ necessary, adjust the constraints on the other entries. */ -+ mp = xmalloc (sizeof (*mp)); -+ mp->fix_size = fix->fix_size; -+ mp->mode = fix->mode; -+ mp->value = fix->value; -+ mp->refcount = 1; -+ /* Not yet required for a backwards ref. */ -+ mp->min_address = -65536; -+ -+ if (max_mp == NULL) -+ { -+ mp->max_address = max_address; -+ mp->next = NULL; -+ mp->prev = minipool_vector_tail; -+ -+ if (mp->prev == NULL) -+ { -+ minipool_vector_head = mp; -+ minipool_vector_label = gen_label_rtx (); -+ } -+ else -+ mp->prev->next = mp; -+ -+ minipool_vector_tail = mp; -+ } -+ else -+ { -+ if (max_address > max_mp->max_address - mp->fix_size) -+ mp->max_address = max_mp->max_address - mp->fix_size; -+ else -+ mp->max_address = max_address; -+ -+ mp->next = max_mp; -+ mp->prev = max_mp->prev; -+ max_mp->prev = mp; -+ if (mp->prev != NULL) -+ mp->prev->next = mp; -+ else -+ minipool_vector_head = mp; -+ } -+ -+ /* Save the new entry. */ -+ max_mp = mp; -+ -+ /* Scan over the preceding entries and adjust their addresses as required. -+ */ -+ while (mp->prev != NULL -+ && mp->prev->max_address > mp->max_address - mp->prev->fix_size) -+ { -+ mp->prev->max_address = mp->max_address - mp->prev->fix_size; -+ mp = mp->prev; -+ } -+ -+ return max_mp; -+} -+ -+static Mnode * -+move_minipool_fix_backward_ref (Mnode * mp, Mnode * min_mp, -+ HOST_WIDE_INT min_address) -+ { -+ HOST_WIDE_INT offset; -+ -+ /* This should never be true, and the code below assumes these are -+ different. */ -+ if (mp == min_mp) -+ abort (); -+ -+ if (min_mp == NULL) -+ { -+ if (min_address > mp->min_address) -+ mp->min_address = min_address; -+ } -+ else -+ { -+ /* We will adjust this below if it is too loose. */ -+ mp->min_address = min_address; -+ -+ /* Unlink MP from its current position. Since min_mp is non-null, -+ mp->next must be non-null. */ -+ mp->next->prev = mp->prev; -+ if (mp->prev != NULL) -+ mp->prev->next = mp->next; -+ else -+ minipool_vector_head = mp->next; -+ -+ /* Reinsert it after MIN_MP. */ -+ mp->prev = min_mp; -+ mp->next = min_mp->next; -+ min_mp->next = mp; -+ if (mp->next != NULL) -+ mp->next->prev = mp; -+ else -+ minipool_vector_tail = mp; -+ } -+ -+ min_mp = mp; -+ -+ offset = 0; -+ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) -+ { -+ mp->offset = offset; -+ if (mp->refcount > 0) -+ offset += mp->fix_size; -+ -+ if (mp->next && mp->next->min_address < mp->min_address + mp->fix_size) -+ mp->next->min_address = mp->min_address + mp->fix_size; -+ } -+ -+ return min_mp; -+ } -+ -+/* Add a constant to the minipool for a backward reference. Returns the -+ node added or NULL if the constant will not fit in this pool. -+ -+ Note that the code for insertion for a backwards reference can be -+ somewhat confusing because the calculated offsets for each fix do -+ not take into account the size of the pool (which is still under -+ construction. */ -+static Mnode * -+add_minipool_backward_ref (Mfix * fix) -+{ -+ /* If set, min_mp is the last pool_entry that has a lower constraint than -+ the one we are trying to add. */ -+ Mnode *min_mp = NULL; -+ /* This can be negative, since it is only a constraint. */ -+ HOST_WIDE_INT min_address = fix->address - fix->backwards; -+ Mnode *mp; -+ -+ /* If we can't reach the current pool from this insn, or if we can't insert -+ this entry at the end of the pool without pushing other fixes out of -+ range, then we don't try. This ensures that we can't fail later on. */ -+ if (min_address >= minipool_barrier->address -+ || (minipool_vector_tail->min_address + fix->fix_size -+ >= minipool_barrier->address)) -+ return NULL; -+ -+ /* Scan the pool to see if a constant with the same value has already been -+ added. While we are doing this, also note the location where we must -+ insert the constant if it doesn't already exist. */ -+ for (mp = minipool_vector_tail; mp != NULL; mp = mp->prev) -+ { -+ if (GET_CODE (fix->value) == GET_CODE (mp->value) -+ && fix->mode == mp->mode -+ && (GET_CODE (fix->value) != CODE_LABEL -+ || (CODE_LABEL_NUMBER (fix->value) -+ == CODE_LABEL_NUMBER (mp->value))) -+ && rtx_equal_p (fix->value, mp->value) -+ /* Check that there is enough slack to move this entry to the end -+ of the table (this is conservative). */ -+ && (mp->max_address -+ > (minipool_barrier->address -+ + minipool_vector_tail->offset -+ + minipool_vector_tail->fix_size))) -+ { -+ mp->refcount++; -+ return move_minipool_fix_backward_ref (mp, min_mp, min_address); -+ } -+ -+ if (min_mp != NULL) -+ mp->min_address += fix->fix_size; -+ else -+ { -+ /* Note the insertion point if necessary. */ -+ if (mp->min_address < min_address) -+ { -+ min_mp = mp; -+ } -+ else if (mp->max_address -+ < minipool_barrier->address + mp->offset + fix->fix_size) -+ { -+ /* Inserting before this entry would push the fix beyond its -+ maximum address (which can happen if we have re-located a -+ forwards fix); force the new fix to come after it. */ -+ min_mp = mp; -+ min_address = mp->min_address + fix->fix_size; -+ } -+ } -+ } -+ -+ /* We need to create a new entry. */ -+ mp = xmalloc (sizeof (*mp)); -+ mp->fix_size = fix->fix_size; -+ mp->mode = fix->mode; -+ mp->value = fix->value; -+ mp->refcount = 1; -+ mp->max_address = minipool_barrier->address + 65536; -+ -+ mp->min_address = min_address; -+ -+ if (min_mp == NULL) -+ { -+ mp->prev = NULL; -+ mp->next = minipool_vector_head; -+ -+ if (mp->next == NULL) -+ { -+ minipool_vector_tail = mp; -+ minipool_vector_label = gen_label_rtx (); -+ } -+ else -+ mp->next->prev = mp; -+ -+ minipool_vector_head = mp; -+ } -+ else -+ { -+ mp->next = min_mp->next; -+ mp->prev = min_mp; -+ min_mp->next = mp; -+ -+ if (mp->next != NULL) -+ mp->next->prev = mp; -+ else -+ minipool_vector_tail = mp; -+ } -+ -+ /* Save the new entry. */ -+ min_mp = mp; -+ -+ if (mp->prev) -+ mp = mp->prev; -+ else -+ mp->offset = 0; -+ -+ /* Scan over the following entries and adjust their offsets. */ -+ while (mp->next != NULL) -+ { -+ if (mp->next->min_address < mp->min_address + mp->fix_size) -+ mp->next->min_address = mp->min_address + mp->fix_size; -+ -+ if (mp->refcount) -+ mp->next->offset = mp->offset + mp->fix_size; -+ else -+ mp->next->offset = mp->offset; -+ -+ mp = mp->next; -+ } -+ -+ return min_mp; -+} -+ -+static void -+assign_minipool_offsets (Mfix * barrier) -+ { -+ HOST_WIDE_INT offset = 0; -+ Mnode *mp; -+ -+ minipool_barrier = barrier; -+ -+ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) -+ { -+ mp->offset = offset; -+ -+ if (mp->refcount > 0) -+ offset += mp->fix_size; -+ } -+ } -+ -+/* Print a symbolic form of X to the debug file, F. */ -+static void -+avr32_print_value (FILE * f, rtx x) -+ { -+ switch (GET_CODE (x)) -+ { -+ case CONST_INT: -+ fprintf (f, "0x%x", (int) INTVAL (x)); -+ return; -+ -+ case CONST_DOUBLE: -+ fprintf (f, "<0x%lx,0x%lx>", (long) XWINT (x, 2), (long) XWINT (x, 3)); -+ return; -+ -+ case CONST_VECTOR: -+ { -+ int i; -+ -+ fprintf (f, "<"); -+ for (i = 0; i < CONST_VECTOR_NUNITS (x); i++) -+ { -+ fprintf (f, "0x%x", (int) INTVAL (CONST_VECTOR_ELT (x, i))); -+ if (i < (CONST_VECTOR_NUNITS (x) - 1)) -+ fputc (',', f); -+ } -+ fprintf (f, ">"); -+ } -+ return; -+ -+ case CONST_STRING: -+ fprintf (f, "\"%s\"", XSTR (x, 0)); -+ return; -+ -+ case SYMBOL_REF: -+ fprintf (f, "`%s'", XSTR (x, 0)); -+ return; -+ -+ case LABEL_REF: -+ fprintf (f, "L%d", INSN_UID (XEXP (x, 0))); -+ return; -+ -+ case CONST: -+ avr32_print_value (f, XEXP (x, 0)); -+ return; -+ -+ case PLUS: -+ avr32_print_value (f, XEXP (x, 0)); -+ fprintf (f, "+"); -+ avr32_print_value (f, XEXP (x, 1)); -+ return; -+ -+ case PC: -+ fprintf (f, "pc"); -+ return; -+ -+ default: -+ fprintf (f, "????"); -+ return; -+ } -+ } -+ -+int -+is_minipool_label (rtx label) -+ { -+ minipool_labels *cur_mp_label = cfun->machine->minipool_label_head; -+ -+ if (GET_CODE (label) != CODE_LABEL) -+ return FALSE; -+ -+ while (cur_mp_label) -+ { -+ if (CODE_LABEL_NUMBER (label) -+ == CODE_LABEL_NUMBER (cur_mp_label->label)) -+ return TRUE; -+ cur_mp_label = cur_mp_label->next; -+ } -+ return FALSE; -+ } -+ -+static void -+new_minipool_label (rtx label) -+ { -+ if (!cfun->machine->minipool_label_head) -+ { -+ cfun->machine->minipool_label_head = -+ ggc_alloc (sizeof (minipool_labels)); -+ cfun->machine->minipool_label_tail = cfun->machine->minipool_label_head; -+ cfun->machine->minipool_label_head->label = label; -+ cfun->machine->minipool_label_head->next = 0; -+ cfun->machine->minipool_label_head->prev = 0; -+ } -+ else -+ { -+ cfun->machine->minipool_label_tail->next = -+ ggc_alloc (sizeof (minipool_labels)); -+ cfun->machine->minipool_label_tail->next->label = label; -+ cfun->machine->minipool_label_tail->next->next = 0; -+ cfun->machine->minipool_label_tail->next->prev = -+ cfun->machine->minipool_label_tail; -+ cfun->machine->minipool_label_tail = -+ cfun->machine->minipool_label_tail->next; -+ } -+ } -+ -+/* Output the literal table */ -+static void -+dump_minipool (rtx scan) -+ { -+ Mnode *mp; -+ Mnode *nmp; -+ -+ if (dump_file) -+ fprintf (dump_file, -+ ";; Emitting minipool after insn %u; address %ld; align %d (bytes)\n", -+ INSN_UID (scan), (unsigned long) minipool_barrier->address, 4); -+ -+ scan = emit_insn_after (gen_consttable_start (), scan); -+ scan = emit_insn_after (gen_align_4 (), scan); -+ scan = emit_label_after (minipool_vector_label, scan); -+ new_minipool_label (minipool_vector_label); -+ -+ for (mp = minipool_vector_head; mp != NULL; mp = nmp) -+ { -+ if (mp->refcount > 0) -+ { -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; Offset %u, min %ld, max %ld ", -+ (unsigned) mp->offset, (unsigned long) mp->min_address, -+ (unsigned long) mp->max_address); -+ avr32_print_value (dump_file, mp->value); -+ fputc ('\n', dump_file); -+ } -+ -+ switch (mp->fix_size) -+ { -+#ifdef HAVE_consttable_4 -+ case 4: -+ scan = emit_insn_after (gen_consttable_4 (mp->value), scan); -+ break; -+ -+#endif -+#ifdef HAVE_consttable_8 -+ case 8: -+ scan = emit_insn_after (gen_consttable_8 (mp->value), scan); -+ break; -+ -+#endif -+#ifdef HAVE_consttable_16 -+ case 16: -+ scan = emit_insn_after (gen_consttable_16 (mp->value), scan); -+ break; -+ -+#endif -+ case 0: -+ /* This can happen for force-minipool entries which just are -+ there to force the minipool to be generate. */ -+ break; -+ default: -+ abort (); -+ break; -+ } -+ } -+ -+ nmp = mp->next; -+ free (mp); -+ } -+ -+ minipool_vector_head = minipool_vector_tail = NULL; -+ scan = emit_insn_after (gen_consttable_end (), scan); -+ scan = emit_barrier_after (scan); -+ } -+ -+/* Return the cost of forcibly inserting a barrier after INSN. */ -+static int -+avr32_barrier_cost (rtx insn) -+ { -+ /* Basing the location of the pool on the loop depth is preferable, but at -+ the moment, the basic block information seems to be corrupt by this -+ stage of the compilation. */ -+ int base_cost = 50; -+ rtx next = next_nonnote_insn (insn); -+ -+ if (next != NULL && GET_CODE (next) == CODE_LABEL) -+ base_cost -= 20; -+ -+ switch (GET_CODE (insn)) -+ { -+ case CODE_LABEL: -+ /* It will always be better to place the table before the label, rather -+ than after it. */ -+ return 50; -+ -+ case INSN: -+ case CALL_INSN: -+ return base_cost; -+ -+ case JUMP_INSN: -+ return base_cost - 10; -+ -+ default: -+ return base_cost + 10; -+ } -+ } -+ -+/* Find the best place in the insn stream in the range -+ (FIX->address,MAX_ADDRESS) to forcibly insert a minipool barrier. -+ Create the barrier by inserting a jump and add a new fix entry for -+ it. */ -+static Mfix * -+create_fix_barrier (Mfix * fix, HOST_WIDE_INT max_address) -+{ -+ HOST_WIDE_INT count = 0; -+ rtx barrier; -+ rtx from = fix->insn; -+ rtx selected = from; -+ int selected_cost; -+ HOST_WIDE_INT selected_address; -+ Mfix *new_fix; -+ HOST_WIDE_INT max_count = max_address - fix->address; -+ rtx label = gen_label_rtx (); -+ -+ selected_cost = avr32_barrier_cost (from); -+ selected_address = fix->address; -+ -+ while (from && count < max_count) -+ { -+ rtx tmp; -+ int new_cost; -+ -+ /* This code shouldn't have been called if there was a natural barrier -+ within range. */ -+ if (GET_CODE (from) == BARRIER) -+ abort (); -+ -+ /* Count the length of this insn. */ -+ count += get_attr_length (from); -+ -+ /* If there is a jump table, add its length. */ -+ tmp = is_jump_table (from); -+ if (tmp != NULL) -+ { -+ count += get_jump_table_size (tmp); -+ -+ /* Jump tables aren't in a basic block, so base the cost on the -+ dispatch insn. If we select this location, we will still put -+ the pool after the table. */ -+ new_cost = avr32_barrier_cost (from); -+ -+ if (count < max_count && new_cost <= selected_cost) -+ { -+ selected = tmp; -+ selected_cost = new_cost; -+ selected_address = fix->address + count; -+ } -+ -+ /* Continue after the dispatch table. */ -+ from = NEXT_INSN (tmp); -+ continue; -+ } -+ -+ new_cost = avr32_barrier_cost (from); -+ -+ if (count < max_count && new_cost <= selected_cost) -+ { -+ selected = from; -+ selected_cost = new_cost; -+ selected_address = fix->address + count; -+ } -+ -+ from = NEXT_INSN (from); -+ } -+ -+ /* Create a new JUMP_INSN that branches around a barrier. */ -+ from = emit_jump_insn_after (gen_jump (label), selected); -+ JUMP_LABEL (from) = label; -+ barrier = emit_barrier_after (from); -+ emit_label_after (label, barrier); -+ -+ /* Create a minipool barrier entry for the new barrier. */ -+ new_fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*new_fix)); -+ new_fix->insn = barrier; -+ new_fix->address = selected_address; -+ new_fix->next = fix->next; -+ fix->next = new_fix; -+ -+ return new_fix; -+} -+ -+/* Record that there is a natural barrier in the insn stream at -+ ADDRESS. */ -+static void -+push_minipool_barrier (rtx insn, HOST_WIDE_INT address) -+ { -+ Mfix *fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*fix)); -+ -+ fix->insn = insn; -+ fix->address = address; -+ -+ fix->next = NULL; -+ if (minipool_fix_head != NULL) -+ minipool_fix_tail->next = fix; -+ else -+ minipool_fix_head = fix; -+ -+ minipool_fix_tail = fix; -+ } -+ -+/* Record INSN, which will need fixing up to load a value from the -+ minipool. ADDRESS is the offset of the insn since the start of the -+ function; LOC is a pointer to the part of the insn which requires -+ fixing; VALUE is the constant that must be loaded, which is of type -+ MODE. */ -+static void -+push_minipool_fix (rtx insn, HOST_WIDE_INT address, rtx * loc, -+ enum machine_mode mode, rtx value) -+ { -+ Mfix *fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*fix)); -+ rtx body = PATTERN (insn); -+ -+ fix->insn = insn; -+ fix->address = address; -+ fix->loc = loc; -+ fix->mode = mode; -+ fix->fix_size = MINIPOOL_FIX_SIZE (mode, value); -+ fix->value = value; -+ -+ if (GET_CODE (body) == PARALLEL) -+ { -+ /* Mcall : Ks16 << 2 */ -+ fix->forwards = ((1 << 15) - 1) << 2; -+ fix->backwards = (1 << 15) << 2; -+ } -+ else if (GET_CODE (body) == SET -+ && GET_MODE_SIZE (GET_MODE (SET_DEST (body))) == 4) -+ { -+ /* Word Load */ -+ if (TARGET_HARD_FLOAT -+ && GET_MODE_CLASS (GET_MODE (SET_DEST (body))) == MODE_FLOAT) -+ { -+ /* Ldc0.w : Ku12 << 2 */ -+ fix->forwards = ((1 << 12) - 1) << 2; -+ fix->backwards = 0; -+ } -+ else -+ { -+ if (optimize_size) -+ { -+ /* Lddpc : Ku7 << 2 */ -+ fix->forwards = ((1 << 7) - 1) << 2; -+ fix->backwards = 0; -+ } -+ else -+ { -+ /* Ld.w : Ks16 */ -+ fix->forwards = ((1 << 15) - 4); -+ fix->backwards = (1 << 15); -+ } -+ } -+ } -+ else if (GET_CODE (body) == SET -+ && GET_MODE_SIZE (GET_MODE (SET_DEST (body))) == 8) -+ { -+ /* Double word load */ -+ if (TARGET_HARD_FLOAT -+ && GET_MODE_CLASS (GET_MODE (SET_DEST (body))) == MODE_FLOAT) -+ { -+ /* Ldc0.d : Ku12 << 2 */ -+ fix->forwards = ((1 << 12) - 1) << 2; -+ fix->backwards = 0; -+ } -+ else -+ { -+ /* Ld.d : Ks16 */ -+ fix->forwards = ((1 << 15) - 4); -+ fix->backwards = (1 << 15); -+ } -+ } -+ else if (GET_CODE (body) == UNSPEC_VOLATILE -+ && XINT (body, 1) == VUNSPEC_MVRC) -+ { -+ /* Coprocessor load */ -+ /* Ldc : Ku8 << 2 */ -+ fix->forwards = ((1 << 8) - 1) << 2; -+ fix->backwards = 0; -+ } -+ else -+ { -+ /* Assume worst case which is lddpc insn. */ -+ fix->forwards = ((1 << 7) - 1) << 2; -+ fix->backwards = 0; -+ } -+ -+ fix->minipool = NULL; -+ -+ /* If an insn doesn't have a range defined for it, then it isn't expecting -+ to be reworked by this code. Better to abort now than to generate duff -+ assembly code. */ -+ if (fix->forwards == 0 && fix->backwards == 0) -+ abort (); -+ -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; %smode fixup for i%d; addr %lu, range (%ld,%ld): ", -+ GET_MODE_NAME (mode), -+ INSN_UID (insn), (unsigned long) address, -+ -1 * (long) fix->backwards, (long) fix->forwards); -+ avr32_print_value (dump_file, fix->value); -+ fprintf (dump_file, "\n"); -+ } -+ -+ /* Add it to the chain of fixes. */ -+ fix->next = NULL; -+ -+ if (minipool_fix_head != NULL) -+ minipool_fix_tail->next = fix; -+ else -+ minipool_fix_head = fix; -+ -+ minipool_fix_tail = fix; -+ } -+ -+/* Scan INSN and note any of its operands that need fixing. -+ If DO_PUSHES is false we do not actually push any of the fixups -+ needed. The function returns TRUE is any fixups were needed/pushed. -+ This is used by avr32_memory_load_p() which needs to know about loads -+ of constants that will be converted into minipool loads. */ -+static bool -+note_invalid_constants (rtx insn, HOST_WIDE_INT address, int do_pushes) -+ { -+ bool result = false; -+ int opno; -+ -+ extract_insn (insn); -+ -+ if (!constrain_operands (1)) -+ fatal_insn_not_found (insn); -+ -+ if (recog_data.n_alternatives == 0) -+ return false; -+ -+ /* Fill in recog_op_alt with information about the constraints of this -+ insn. */ -+ preprocess_constraints (); -+ -+ for (opno = 0; opno < recog_data.n_operands; opno++) -+ { -+ rtx op; -+ -+ /* Things we need to fix can only occur in inputs. */ -+ if (recog_data.operand_type[opno] != OP_IN) -+ continue; -+ -+ op = recog_data.operand[opno]; -+ -+ if (avr32_const_pool_ref_operand (op, GET_MODE (op))) -+ { -+ if (do_pushes) -+ { -+ rtx cop = avoid_constant_pool_reference (op); -+ -+ /* Casting the address of something to a mode narrower than a -+ word can cause avoid_constant_pool_reference() to return the -+ pool reference itself. That's no good to us here. Lets -+ just hope that we can use the constant pool value directly. -+ */ -+ if (op == cop) -+ cop = get_pool_constant (XEXP (op, 0)); -+ -+ push_minipool_fix (insn, address, -+ recog_data.operand_loc[opno], -+ recog_data.operand_mode[opno], cop); -+ } -+ -+ result = true; -+ } -+ else if (TARGET_HAS_ASM_ADDR_PSEUDOS -+ && avr32_address_operand (op, GET_MODE (op))) -+ { -+ /* Handle pseudo instructions using a direct address. These pseudo -+ instructions might need entries in the constant pool and we must -+ therefor create a constant pool for them, in case the -+ assembler/linker needs to insert entries. */ -+ if (do_pushes) -+ { -+ /* Push a dummy constant pool entry so that the .cpool -+ directive should be inserted on the appropriate place in the -+ code even if there are no real constant pool entries. This -+ is used by the assembler and linker to know where to put -+ generated constant pool entries. */ -+ push_minipool_fix (insn, address, -+ recog_data.operand_loc[opno], -+ recog_data.operand_mode[opno], -+ gen_rtx_UNSPEC (VOIDmode, -+ gen_rtvec (1, const0_rtx), -+ UNSPEC_FORCE_MINIPOOL)); -+ result = true; -+ } -+ } -+ } -+ return result; -+ } -+ -+ -+static int -+avr32_insn_is_cast (rtx insn) -+ { -+ -+ if (NONJUMP_INSN_P (insn) -+ && GET_CODE (PATTERN (insn)) == SET -+ && (GET_CODE (SET_SRC (PATTERN (insn))) == ZERO_EXTEND -+ || GET_CODE (SET_SRC (PATTERN (insn))) == SIGN_EXTEND) -+ && REG_P (XEXP (SET_SRC (PATTERN (insn)), 0)) -+ && REG_P (SET_DEST (PATTERN (insn)))) -+ return true; -+ return false; -+ } -+ -+/* FIXME: The level of nesting in this function is way too deep. It needs to be -+ torn apart. */ -+static void -+avr32_reorg_optimization (void) -+ { -+ rtx first = get_insns (); -+ rtx insn; -+ -+ if (TARGET_MD_REORG_OPTIMIZATION && (optimize_size || (optimize > 0))) -+ { -+ -+ /* Scan through all insns looking for cast operations. */ -+ if (dump_file) -+ { -+ fprintf (dump_file, ";; Deleting redundant cast operations:\n"); -+ } -+ for (insn = first; insn; insn = NEXT_INSN (insn)) -+ { -+ rtx reg, src_reg, scan; -+ enum machine_mode mode; -+ int unused_cast; -+ rtx label_ref; -+ -+ if (avr32_insn_is_cast (insn) -+ && (GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == QImode -+ || GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == HImode)) -+ { -+ mode = GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)); -+ reg = SET_DEST (PATTERN (insn)); -+ src_reg = XEXP (SET_SRC (PATTERN (insn)), 0); -+ } -+ else -+ { -+ continue; -+ } -+ -+ unused_cast = false; -+ label_ref = NULL_RTX; -+ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) -+ { -+ /* Check if we have reached the destination of a simple -+ conditional jump which we have already scanned past. If so, -+ we can safely continue scanning. */ -+ if (LABEL_P (scan) && label_ref != NULL_RTX) -+ { -+ if (CODE_LABEL_NUMBER (scan) == -+ CODE_LABEL_NUMBER (XEXP (label_ref, 0))) -+ label_ref = NULL_RTX; -+ else -+ break; -+ } -+ -+ if (!INSN_P (scan)) -+ continue; -+ -+ /* For conditional jumps we can manage to keep on scanning if -+ we meet the destination label later on before any new jump -+ insns occure. */ -+ if (GET_CODE (scan) == JUMP_INSN) -+ { -+ if (any_condjump_p (scan) && label_ref == NULL_RTX) -+ label_ref = condjump_label (scan); -+ else -+ break; -+ } -+ -+ if (!reg_mentioned_p (reg, PATTERN (scan))) -+ continue; -+ -+ /* Check if casted register is used in this insn */ -+ if ((regno_use_in (REGNO (reg), PATTERN (scan)) != NULL_RTX) -+ && (GET_MODE (regno_use_in (REGNO (reg), PATTERN (scan))) == -+ GET_MODE (reg))) -+ { -+ /* If not used in the source to the set or in a memory -+ expression in the destiantion then the register is used -+ as a destination and is really dead. */ -+ if (single_set (scan) -+ && GET_CODE (PATTERN (scan)) == SET -+ && REG_P (SET_DEST (PATTERN (scan))) -+ && !regno_use_in (REGNO (reg), SET_SRC (PATTERN (scan))) -+ && label_ref == NULL_RTX) -+ { -+ unused_cast = true; -+ } -+ break; -+ } -+ -+ /* Check if register is dead or set in this insn */ -+ if (dead_or_set_p (scan, reg)) -+ { -+ unused_cast = true; -+ break; -+ } -+ } -+ -+ /* Check if we have unresolved conditional jumps */ -+ if (label_ref != NULL_RTX) -+ continue; -+ -+ if (unused_cast) -+ { -+ if (REGNO (reg) == REGNO (XEXP (SET_SRC (PATTERN (insn)), 0))) -+ { -+ /* One operand cast, safe to delete */ -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; INSN %i removed, casted register %i value not used.\n", -+ INSN_UID (insn), REGNO (reg)); -+ } -+ SET_INSN_DELETED (insn); -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (insn) = -1; -+ } -+ else -+ { -+ /* Two operand cast, which really could be substituted with -+ a move, if the source register is dead after the cast -+ insn and then the insn which sets the source register -+ could instead directly set the destination register for -+ the cast. As long as there are no insns in between which -+ uses the register. */ -+ rtx link = NULL_RTX; -+ rtx set; -+ rtx src_reg = XEXP (SET_SRC (PATTERN (insn)), 0); -+ unused_cast = false; -+ -+ if (!find_reg_note (insn, REG_DEAD, src_reg)) -+ continue; -+ -+ /* Search for the insn which sets the source register */ -+ for (link = LOG_LINKS (insn); link; link = XEXP (link, 1)) -+ { -+ if (REG_NOTE_KIND (link) != 0) -+ continue; -+ set = single_set (XEXP (link, 0)); -+ if (set && rtx_equal_p (src_reg, SET_DEST (set))) -+ { -+ link = XEXP (link, 0); -+ break; -+ } -+ } -+ -+ /* Found no link or link is a call insn where we can not -+ change the destination register */ -+ if (link == NULL_RTX || CALL_P (link)) -+ continue; -+ -+ /* Scan through all insn between link and insn */ -+ for (scan = NEXT_INSN (link); scan; scan = NEXT_INSN (scan)) -+ { -+ /* Don't try to trace forward past a CODE_LABEL if we -+ haven't seen INSN yet. Ordinarily, we will only -+ find the setting insn in LOG_LINKS if it is in the -+ same basic block. However, cross-jumping can insert -+ code labels in between the load and the call, and -+ can result in situations where a single call insn -+ may have two targets depending on where we came -+ from. */ -+ -+ if (GET_CODE (scan) == CODE_LABEL) -+ break; -+ -+ if (!INSN_P (scan)) -+ continue; -+ -+ /* Don't try to trace forward past a JUMP. To optimize -+ safely, we would have to check that all the -+ instructions at the jump destination did not use REG. -+ */ -+ -+ if (GET_CODE (scan) == JUMP_INSN) -+ { -+ break; -+ } -+ -+ if (!reg_mentioned_p (src_reg, PATTERN (scan))) -+ continue; -+ -+ /* We have reached the cast insn */ -+ if (scan == insn) -+ { -+ /* We can remove cast and replace the destination -+ register of the link insn with the destination -+ of the cast */ -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; INSN %i removed, casted value unused. " -+ "Destination of removed cast operation: register %i, folded into INSN %i.\n", -+ INSN_UID (insn), REGNO (reg), -+ INSN_UID (link)); -+ } -+ /* Update link insn */ -+ SET_DEST (PATTERN (link)) = -+ gen_rtx_REG (mode, REGNO (reg)); -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (link) = -1; -+ -+ /* Delete insn */ -+ SET_INSN_DELETED (insn); -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (insn) = -1; -+ break; -+ } -+ } -+ } -+ } -+ } -+ } -+ -+ if (TARGET_MD_REORG_OPTIMIZATION && (optimize_size || (optimize > 0))) -+ { -+ -+ /* Scan through all insns looking for shifted add operations */ -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; Deleting redundant shifted add operations:\n"); -+ } -+ for (insn = first; insn; insn = NEXT_INSN (insn)) -+ { -+ rtx reg, mem_expr, scan, op0, op1; -+ int add_only_used_as_pointer; -+ -+ if (INSN_P (insn) -+ && GET_CODE (PATTERN (insn)) == SET -+ && GET_CODE (SET_SRC (PATTERN (insn))) == PLUS -+ && (GET_CODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == MULT -+ || GET_CODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == ASHIFT) -+ && GET_CODE (XEXP (XEXP (SET_SRC (PATTERN (insn)), 0), 1)) == -+ CONST_INT && REG_P (SET_DEST (PATTERN (insn))) -+ && REG_P (XEXP (SET_SRC (PATTERN (insn)), 1)) -+ && REG_P (XEXP (XEXP (SET_SRC (PATTERN (insn)), 0), 0))) -+ { -+ reg = SET_DEST (PATTERN (insn)); -+ mem_expr = SET_SRC (PATTERN (insn)); -+ op0 = XEXP (XEXP (mem_expr, 0), 0); -+ op1 = XEXP (mem_expr, 1); -+ } -+ else -+ { -+ continue; -+ } -+ -+ /* Scan forward the check if the result of the shifted add -+ operation is only used as an address in memory operations and -+ that the operands to the shifted add are not clobbered. */ -+ add_only_used_as_pointer = false; -+ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) -+ { -+ if (!INSN_P (scan)) -+ continue; -+ -+ /* Don't try to trace forward past a JUMP or CALL. To optimize -+ safely, we would have to check that all the instructions at -+ the jump destination did not use REG. */ -+ -+ if (GET_CODE (scan) == JUMP_INSN) -+ { -+ break; -+ } -+ -+ /* If used in a call insn then we cannot optimize it away */ -+ if (CALL_P (scan) && find_regno_fusage (scan, USE, REGNO (reg))) -+ break; -+ -+ /* If any of the operands of the shifted add are clobbered we -+ cannot optimize the shifted adda away */ -+ if ((reg_set_p (op0, scan) && (REGNO (op0) != REGNO (reg))) -+ || (reg_set_p (op1, scan) && (REGNO (op1) != REGNO (reg)))) -+ break; -+ -+ if (!reg_mentioned_p (reg, PATTERN (scan))) -+ continue; -+ -+ /* If used any other place than as a pointer or as the -+ destination register we failed */ -+ if (!(single_set (scan) -+ && GET_CODE (PATTERN (scan)) == SET -+ && ((MEM_P (SET_DEST (PATTERN (scan))) -+ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) -+ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == -+ REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) -+ && -+ REG_P (XEXP -+ (SET_SRC (PATTERN (scan)), -+ 0)) -+ && -+ REGNO (XEXP -+ (SET_SRC (PATTERN (scan)), -+ 0)) == REGNO (reg)))) -+ && !(GET_CODE (PATTERN (scan)) == SET -+ && REG_P (SET_DEST (PATTERN (scan))) -+ && !regno_use_in (REGNO (reg), -+ SET_SRC (PATTERN (scan))))) -+ break; -+ -+ /* Check if register is dead or set in this insn */ -+ if (dead_or_set_p (scan, reg)) -+ { -+ add_only_used_as_pointer = true; -+ break; -+ } -+ } -+ -+ if (add_only_used_as_pointer) -+ { -+ /* Lets delete the add insn and replace all memory references -+ which uses the pointer with the full expression. */ -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; Deleting INSN %i since address expression can be folded into all " -+ "memory references using this expression\n", -+ INSN_UID (insn)); -+ } -+ SET_INSN_DELETED (insn); -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (insn) = -1; -+ -+ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) -+ { -+ if (!INSN_P (scan)) -+ continue; -+ -+ if (!reg_mentioned_p (reg, PATTERN (scan))) -+ continue; -+ -+ /* If used any other place than as a pointer or as the -+ destination register we failed */ -+ if ((single_set (scan) -+ && GET_CODE (PATTERN (scan)) == SET -+ && ((MEM_P (SET_DEST (PATTERN (scan))) -+ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) -+ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == -+ REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) -+ && -+ REG_P (XEXP -+ (SET_SRC (PATTERN (scan)), -+ 0)) -+ && -+ REGNO (XEXP -+ (SET_SRC (PATTERN (scan)), -+ 0)) == REGNO (reg))))) -+ { -+ if (dump_file) -+ { -+ fprintf (dump_file, -+ ";; Register %i replaced by indexed address in INSN %i\n", -+ REGNO (reg), INSN_UID (scan)); -+ } -+ if (MEM_P (SET_DEST (PATTERN (scan)))) -+ XEXP (SET_DEST (PATTERN (scan)), 0) = mem_expr; -+ else -+ XEXP (SET_SRC (PATTERN (scan)), 0) = mem_expr; -+ } -+ -+ /* Check if register is dead or set in this insn */ -+ if (dead_or_set_p (scan, reg)) -+ { -+ break; -+ } -+ -+ } -+ } -+ } -+ } -+ } -+ -+/* Exported to toplev.c. -+ -+ Do a final pass over the function, just before delayed branch -+ scheduling. */ -+ -+static void -+avr32_reorg (void) -+ { -+ rtx insn; -+ HOST_WIDE_INT address = 0; -+ Mfix *fix; -+ -+ minipool_fix_head = minipool_fix_tail = NULL; -+ -+ /* The first insn must always be a note, or the code below won't scan it -+ properly. */ -+ insn = get_insns (); -+ if (GET_CODE (insn) != NOTE) -+ abort (); -+ -+ /* Scan all the insns and record the operands that will need fixing. */ -+ for (insn = next_nonnote_insn (insn); insn; insn = next_nonnote_insn (insn)) -+ { -+ if (GET_CODE (insn) == BARRIER) -+ push_minipool_barrier (insn, address); -+ else if (INSN_P (insn)) -+ { -+ rtx table; -+ -+ note_invalid_constants (insn, address, true); -+ address += get_attr_length (insn); -+ -+ /* If the insn is a vector jump, add the size of the table and skip -+ the table. */ -+ if ((table = is_jump_table (insn)) != NULL) -+ { -+ address += get_jump_table_size (table); -+ insn = table; -+ } -+ } -+ } -+ -+ fix = minipool_fix_head; -+ -+ /* Now scan the fixups and perform the required changes. */ -+ while (fix) -+ { -+ Mfix *ftmp; -+ Mfix *fdel; -+ Mfix *last_added_fix; -+ Mfix *last_barrier = NULL; -+ Mfix *this_fix; -+ -+ /* Skip any further barriers before the next fix. */ -+ while (fix && GET_CODE (fix->insn) == BARRIER) -+ fix = fix->next; -+ -+ /* No more fixes. */ -+ if (fix == NULL) -+ break; -+ -+ last_added_fix = NULL; -+ -+ for (ftmp = fix; ftmp; ftmp = ftmp->next) -+ { -+ if (GET_CODE (ftmp->insn) == BARRIER) -+ { -+ if (ftmp->address >= minipool_vector_head->max_address) -+ break; -+ -+ last_barrier = ftmp; -+ } -+ else if ((ftmp->minipool = add_minipool_forward_ref (ftmp)) == NULL) -+ break; -+ -+ last_added_fix = ftmp; /* Keep track of the last fix added. -+ */ -+ } -+ -+ /* If we found a barrier, drop back to that; any fixes that we could -+ have reached but come after the barrier will now go in the next -+ mini-pool. */ -+ if (last_barrier != NULL) -+ { -+ /* Reduce the refcount for those fixes that won't go into this pool -+ after all. */ -+ for (fdel = last_barrier->next; -+ fdel && fdel != ftmp; fdel = fdel->next) -+ { -+ fdel->minipool->refcount--; -+ fdel->minipool = NULL; -+ } -+ -+ ftmp = last_barrier; -+ } -+ else -+ { -+ /* ftmp is first fix that we can't fit into this pool and there no -+ natural barriers that we could use. Insert a new barrier in the -+ code somewhere between the previous fix and this one, and -+ arrange to jump around it. */ -+ HOST_WIDE_INT max_address; -+ -+ /* The last item on the list of fixes must be a barrier, so we can -+ never run off the end of the list of fixes without last_barrier -+ being set. */ -+ if (ftmp == NULL) -+ abort (); -+ -+ max_address = minipool_vector_head->max_address; -+ /* Check that there isn't another fix that is in range that we -+ couldn't fit into this pool because the pool was already too -+ large: we need to put the pool before such an instruction. */ -+ if (ftmp->address < max_address) -+ max_address = ftmp->address; -+ -+ last_barrier = create_fix_barrier (last_added_fix, max_address); -+ } -+ -+ assign_minipool_offsets (last_barrier); -+ -+ while (ftmp) -+ { -+ if (GET_CODE (ftmp->insn) != BARRIER -+ && ((ftmp->minipool = add_minipool_backward_ref (ftmp)) -+ == NULL)) -+ break; -+ -+ ftmp = ftmp->next; -+ } -+ -+ /* Scan over the fixes we have identified for this pool, fixing them up -+ and adding the constants to the pool itself. */ -+ for (this_fix = fix; this_fix && ftmp != this_fix; -+ this_fix = this_fix->next) -+ if (GET_CODE (this_fix->insn) != BARRIER -+ /* Do nothing for entries present just to force the insertion of -+ a minipool. */ -+ && !IS_FORCE_MINIPOOL (this_fix->value)) -+ { -+ rtx addr = plus_constant (gen_rtx_LABEL_REF (VOIDmode, -+ minipool_vector_label), -+ this_fix->minipool->offset); -+ *this_fix->loc = gen_rtx_MEM (this_fix->mode, addr); -+ } -+ -+ dump_minipool (last_barrier->insn); -+ fix = ftmp; -+ } -+ -+ /* Free the minipool memory. */ -+ obstack_free (&minipool_obstack, minipool_startobj); -+ -+ avr32_reorg_optimization (); -+ } -+ -+ -+/* -+ Hook for doing some final scanning of instructions. Does nothing yet...*/ -+void -+avr32_final_prescan_insn (rtx insn ATTRIBUTE_UNUSED, -+ rtx * opvec ATTRIBUTE_UNUSED, -+ int noperands ATTRIBUTE_UNUSED) -+ { -+ return; -+ } -+ -+ -+/* Function for changing the condition on the next instruction, -+ should be used when emmiting compare instructions and -+ the condition of the next instruction needs to change. -+ */ -+int -+set_next_insn_cond (rtx cur_insn, rtx new_cond) -+ { -+ rtx next_insn = next_nonnote_insn (cur_insn); -+ if ((next_insn != NULL_RTX) -+ && (INSN_P (next_insn)) -+ && (GET_CODE (PATTERN (next_insn)) == SET) -+ && (GET_CODE (SET_SRC (PATTERN (next_insn))) == IF_THEN_ELSE)) -+ { -+ /* Branch instructions */ -+ XEXP (SET_SRC (PATTERN (next_insn)), 0) = new_cond; -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (next_insn) = -1; -+ return TRUE; -+ } -+ else if ((next_insn != NULL_RTX) -+ && (INSN_P (next_insn)) -+ && (GET_CODE (PATTERN (next_insn)) == SET) -+ && comparison_operator (SET_SRC (PATTERN (next_insn)), -+ GET_MODE (SET_SRC (PATTERN (next_insn))))) -+ { -+ /* scc with no compare */ -+ SET_SRC (PATTERN (next_insn)) = new_cond; -+ /* Force the instruction to be recognized again */ -+ INSN_CODE (next_insn) = -1; -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+/* Function for obtaining the condition for the next instruction -+ after cur_insn. -+ */ -+rtx -+get_next_insn_cond (rtx cur_insn) -+ { -+ rtx next_insn = next_nonnote_insn (cur_insn); -+ rtx cond = NULL_RTX; -+ if ((next_insn != NULL_RTX) -+ && (INSN_P (next_insn)) -+ && (GET_CODE (PATTERN (next_insn)) == SET) -+ && (GET_CODE (SET_SRC (PATTERN (next_insn))) == IF_THEN_ELSE)) -+ { -+ /* Branch instructions */ -+ cond = XEXP (SET_SRC (PATTERN (next_insn)), 0); -+ } -+ else if ((next_insn != NULL_RTX) -+ && (INSN_P (next_insn)) -+ && (GET_CODE (PATTERN (next_insn)) == SET) -+ && comparison_operator (SET_SRC (PATTERN (next_insn)), -+ GET_MODE (SET_SRC (PATTERN (next_insn))))) -+ { -+ /* scc with no compare */ -+ cond = SET_SRC (PATTERN (next_insn)); -+ } -+ -+ return cond; -+ } -+ -+ -+rtx -+avr32_output_cmp (rtx cond, enum machine_mode mode, rtx op0, rtx op1) -+ { -+ -+ rtx new_cond = NULL_RTX; -+ rtx ops[2]; -+ rtx compare_pattern; -+ ops[0] = op0; -+ ops[1] = op1; -+ -+ compare_pattern = gen_rtx_COMPARE (mode, op0, op1); -+ -+ new_cond = is_compare_redundant (compare_pattern, cond); -+ -+ if (new_cond != NULL_RTX) -+ return new_cond; -+ -+ /* Insert compare */ -+ switch (mode) -+ { -+ case QImode: -+ output_asm_insn ("cp.b\t%0, %1", ops); -+ break; -+ case HImode: -+ output_asm_insn ("cp.h\t%0, %1", ops); -+ break; -+ case SImode: -+ output_asm_insn ("cp.w\t%0, %1", ops); -+ break; -+ case DImode: -+ if (GET_CODE (op1) != REG) -+ output_asm_insn ("cp.w\t%0, %1\ncpc\t%m0", ops); -+ else -+ output_asm_insn ("cp.w\t%0, %1\ncpc\t%m0, %m1", ops); -+ break; -+ default: -+ internal_error ("Unknown comparison mode"); -+ break; -+ } -+ -+ return cond; -+ } -+ -+int -+avr32_load_multiple_operation (rtx op, -+ enum machine_mode mode ATTRIBUTE_UNUSED) -+ { -+ int count = XVECLEN (op, 0); -+ unsigned int dest_regno; -+ rtx src_addr; -+ rtx elt; -+ int i = 1, base = 0; -+ -+ if (count <= 1 || GET_CODE (XVECEXP (op, 0, 0)) != SET) -+ return 0; -+ -+ /* Check to see if this might be a write-back. */ -+ if (GET_CODE (SET_SRC (elt = XVECEXP (op, 0, 0))) == PLUS) -+ { -+ i++; -+ base = 1; -+ -+ /* Now check it more carefully. */ -+ if (GET_CODE (SET_DEST (elt)) != REG -+ || GET_CODE (XEXP (SET_SRC (elt), 0)) != REG -+ || GET_CODE (XEXP (SET_SRC (elt), 1)) != CONST_INT -+ || INTVAL (XEXP (SET_SRC (elt), 1)) != (count - 1) * 4) -+ return 0; -+ } -+ -+ /* Perform a quick check so we don't blow up below. */ -+ if (count <= 1 -+ || GET_CODE (XVECEXP (op, 0, i - 1)) != SET -+ || GET_CODE (SET_DEST (XVECEXP (op, 0, i - 1))) != REG -+ || GET_CODE (SET_SRC (XVECEXP (op, 0, i - 1))) != UNSPEC) -+ return 0; -+ -+ dest_regno = REGNO (SET_DEST (XVECEXP (op, 0, i - 1))); -+ src_addr = XEXP (SET_SRC (XVECEXP (op, 0, i - 1)), 0); -+ -+ for (; i < count; i++) -+ { -+ elt = XVECEXP (op, 0, i); -+ -+ if (GET_CODE (elt) != SET -+ || GET_CODE (SET_DEST (elt)) != REG -+ || GET_MODE (SET_DEST (elt)) != SImode -+ || GET_CODE (SET_SRC (elt)) != UNSPEC) -+ return 0; -+ } -+ -+ return 1; -+ } -+ -+int -+avr32_store_multiple_operation (rtx op, -+ enum machine_mode mode ATTRIBUTE_UNUSED) -+ { -+ int count = XVECLEN (op, 0); -+ int src_regno; -+ rtx dest_addr; -+ rtx elt; -+ int i = 1; -+ -+ if (count <= 1 || GET_CODE (XVECEXP (op, 0, 0)) != SET) -+ return 0; -+ -+ /* Perform a quick check so we don't blow up below. */ -+ if (count <= i -+ || GET_CODE (XVECEXP (op, 0, i - 1)) != SET -+ || GET_CODE (SET_DEST (XVECEXP (op, 0, i - 1))) != MEM -+ || GET_CODE (SET_SRC (XVECEXP (op, 0, i - 1))) != UNSPEC) -+ return 0; -+ -+ src_regno = REGNO (SET_SRC (XVECEXP (op, 0, i - 1))); -+ dest_addr = XEXP (SET_DEST (XVECEXP (op, 0, i - 1)), 0); -+ -+ for (; i < count; i++) -+ { -+ elt = XVECEXP (op, 0, i); -+ -+ if (GET_CODE (elt) != SET -+ || GET_CODE (SET_DEST (elt)) != MEM -+ || GET_MODE (SET_DEST (elt)) != SImode -+ || GET_CODE (SET_SRC (elt)) != UNSPEC) -+ return 0; -+ } -+ -+ return 1; -+ } -+ -+int -+avr32_valid_macmac_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* Check if they use the same accumulator */ -+ if (rtx_equal_p -+ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) -+ { -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+int -+avr32_valid_mulmac_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* -+ Check if the mul instruction produces the accumulator for the mac -+ instruction. */ -+ if (rtx_equal_p -+ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) -+ { -+ return TRUE; -+ } -+ return FALSE; -+ } -+ -+int -+avr32_store_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* Only valid bypass if the output result is used as an src in the store -+ instruction, NOT if used as a pointer or base. */ -+ if (rtx_equal_p -+ (SET_DEST (PATTERN (insn_out)), SET_SRC (PATTERN (insn_in)))) -+ { -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+int -+avr32_mul_waw_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* Check if the register holding the result from the mul instruction is -+ used as a result register in the input instruction. */ -+ if (rtx_equal_p -+ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) -+ { -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+int -+avr32_valid_load_double_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* Check if the first loaded word in insn_out is used in insn_in. */ -+ rtx dst_reg; -+ rtx second_loaded_reg; -+ -+ /* If this is a double alu operation then the bypass is not valid */ -+ if ((get_attr_type (insn_in) == TYPE_ALU -+ || get_attr_type (insn_in) == TYPE_ALU2) -+ && (GET_MODE_SIZE (GET_MODE (SET_DEST (PATTERN (insn_out)))) > 4)) -+ return FALSE; -+ -+ /* Get the destination register in the load */ -+ if (!REG_P (SET_DEST (PATTERN (insn_out)))) -+ return FALSE; -+ -+ dst_reg = SET_DEST (PATTERN (insn_out)); -+ second_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 1); -+ -+ if (!reg_mentioned_p (second_loaded_reg, PATTERN (insn_in))) -+ return TRUE; -+ -+ return FALSE; -+ } -+ -+ -+int -+avr32_valid_load_quad_bypass (rtx insn_out, rtx insn_in) -+ { -+ /* -+ Check if the two first loaded word in insn_out are used in insn_in. */ -+ rtx dst_reg; -+ rtx third_loaded_reg, fourth_loaded_reg; -+ -+ /* Get the destination register in the load */ -+ if (!REG_P (SET_DEST (PATTERN (insn_out)))) -+ return FALSE; -+ -+ dst_reg = SET_DEST (PATTERN (insn_out)); -+ third_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 2); -+ fourth_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 3); -+ -+ if (!reg_mentioned_p (third_loaded_reg, PATTERN (insn_in)) -+ && !reg_mentioned_p (fourth_loaded_reg, PATTERN (insn_in))) -+ { -+ return TRUE; -+ } -+ -+ return FALSE; -+ } -+ -+ -+//section * -+//avr32_select_rtx_section (enum machine_mode mode ATTRIBUTE_UNUSED, -+// rtx x ATTRIBUTE_UNUSED, -+// unsigned HOST_WIDE_INT align ATTRIBUTE_UNUSED) -+// { -+// /* Let ASM_OUTPUT_POOL_PROLOGUE take care of this */ -+// return 0; -+// } -+ -+ -+/* Function for getting an integer value from a const_int or const_double -+ expression regardless of the HOST_WIDE_INT size. Each target cpu word -+ will be put into the val array where the LSW will be stored at the lowest -+ address and so forth. Assumes that const_expr is either a const_int or -+ const_double. Only valid for modes which have sizes that are a multiple -+ of the word size. -+*/ -+void -+avr32_get_intval (enum machine_mode mode, -+ rtx const_expr, -+ HOST_WIDE_INT *val) -+{ -+ int words_in_mode = GET_MODE_SIZE (mode)/UNITS_PER_WORD; -+ const int words_in_const_int = HOST_BITS_PER_WIDE_INT / BITS_PER_WORD; -+ -+ if ( GET_CODE(const_expr) == CONST_DOUBLE ){ -+ HOST_WIDE_INT hi = CONST_DOUBLE_HIGH(const_expr); -+ HOST_WIDE_INT lo = CONST_DOUBLE_LOW(const_expr); -+ /* Evaluate hi and lo values of const_double. */ -+ avr32_get_intval (mode_for_size (HOST_BITS_PER_WIDE_INT, MODE_INT, 0), -+ GEN_INT (lo), -+ &val[0]); -+ avr32_get_intval (mode_for_size (HOST_BITS_PER_WIDE_INT, MODE_INT, 0), -+ GEN_INT (hi), -+ &val[words_in_const_int]); -+ } else if ( GET_CODE(const_expr) == CONST_INT ){ -+ HOST_WIDE_INT value = INTVAL(const_expr); -+ int word; -+ for ( word = 0; (word < words_in_mode) && (word < words_in_const_int); word++ ){ -+ /* Shift word up to the MSW and shift down again to extract the -+ word and sign-extend. */ -+ int lshift = (words_in_const_int - word - 1) * BITS_PER_WORD; -+ int rshift = (words_in_const_int-1) * BITS_PER_WORD; -+ val[word] = (value << lshift) >> rshift; -+ } -+ -+ for ( ; word < words_in_mode; word++ ){ -+ /* Just put the sign bits in the remaining words. */ -+ val[word] = value < 0 ? -1 : 0; -+ } -+ } -+} -+ -+void -+avr32_split_const_expr (enum machine_mode mode, -+ enum machine_mode new_mode, -+ rtx expr, -+ rtx *split_expr) -+{ -+ int i, word; -+ int words_in_intval = GET_MODE_SIZE (mode)/UNITS_PER_WORD; -+ int words_in_split_values = GET_MODE_SIZE (new_mode)/UNITS_PER_WORD; -+ const int words_in_const_int = HOST_BITS_PER_WIDE_INT / BITS_PER_WORD; -+ HOST_WIDE_INT *val = alloca (words_in_intval * UNITS_PER_WORD); -+ -+ avr32_get_intval (mode, expr, val); -+ -+ for ( i=0; i < (words_in_intval/words_in_split_values); i++ ) -+ { -+ HOST_WIDE_INT value_lo = 0, value_hi = 0; -+ for ( word = 0; word < words_in_split_values; word++ ) -+ { -+ if ( word >= words_in_const_int ) -+ value_hi |= ((val[i * words_in_split_values + word] & -+ (((HOST_WIDE_INT)1 << BITS_PER_WORD)-1)) -+ << (BITS_PER_WORD * (word - words_in_const_int))); -+ else -+ value_lo |= ((val[i * words_in_split_values + word] & -+ (((HOST_WIDE_INT)1 << BITS_PER_WORD)-1)) -+ << (BITS_PER_WORD * word)); -+ } -+ split_expr[i] = immed_double_const(value_lo, value_hi, new_mode); -+ } -+} -+ -+ -+ -+/* Set up library functions to comply to AVR32 ABI */ -+ -+ -+/* Set up library functions to comply to AVR32 ABI */ -+ -+static void -+avr32_init_libfuncs (void) -+{ -+ /* Convert gcc run-time function names to AVR32 ABI names */ -+ -+ /* Double-precision floating-point arithmetic. */ -+ set_optab_libfunc (sdiv_optab, DFmode, "__avr32_f64_div"); -+ set_optab_libfunc (smul_optab, DFmode, "__avr32_f64_mul"); -+ set_optab_libfunc (neg_optab, DFmode, NULL); -+ -+ /* Double-precision comparisons. */ -+ set_optab_libfunc (eq_optab, DFmode, "__avr32_f64_cmp_eq"); -+ set_optab_libfunc (ne_optab, DFmode, NULL); -+ set_optab_libfunc (lt_optab, DFmode, "__avr32_f64_cmp_lt"); -+ set_optab_libfunc (le_optab, DFmode, NULL); -+ set_optab_libfunc (ge_optab, DFmode, "__avr32_f64_cmp_ge"); -+ set_optab_libfunc (gt_optab, DFmode, NULL); -+ -+ /* Single-precision floating-point arithmetic. */ -+ set_optab_libfunc (smul_optab, SFmode, "__avr32_f32_mul"); -+ set_optab_libfunc (neg_optab, SFmode, NULL); -+ -+ /* Single-precision comparisons. */ -+ set_optab_libfunc (eq_optab, SFmode, "__avr32_f32_cmp_eq"); -+ set_optab_libfunc (ne_optab, SFmode, NULL); -+ set_optab_libfunc (lt_optab, SFmode, "__avr32_f32_cmp_lt"); -+ set_optab_libfunc (le_optab, SFmode, NULL); -+ set_optab_libfunc (ge_optab, SFmode, "__avr32_f32_cmp_ge"); -+ set_optab_libfunc (gt_optab, SFmode, NULL); -+ -+ /* Floating-point to integer conversions. */ -+ set_conv_libfunc (sfix_optab, SImode, DFmode, "__avr32_f64_to_s32"); -+ set_conv_libfunc (ufix_optab, SImode, DFmode, "__avr32_f64_to_u32"); -+ set_conv_libfunc (sfix_optab, DImode, DFmode, "__avr32_f64_to_s64"); -+ set_conv_libfunc (ufix_optab, DImode, DFmode, "__avr32_f64_to_u64"); -+ set_conv_libfunc (sfix_optab, SImode, SFmode, "__avr32_f32_to_s32"); -+ set_conv_libfunc (ufix_optab, SImode, SFmode, "__avr32_f32_to_u32"); -+ set_conv_libfunc (sfix_optab, DImode, SFmode, "__avr32_f32_to_s64"); -+ set_conv_libfunc (ufix_optab, DImode, SFmode, "__avr32_f32_to_u64"); -+ -+ /* Conversions between floating types. */ -+ set_conv_libfunc (trunc_optab, SFmode, DFmode, "__avr32_f64_to_f32"); -+ set_conv_libfunc (sext_optab, DFmode, SFmode, "__avr32_f32_to_f64"); -+ -+ /* Integer to floating-point conversions. Table 8. */ -+ set_conv_libfunc (sfloat_optab, DFmode, SImode, "__avr32_s32_to_f64"); -+ set_conv_libfunc (sfloat_optab, DFmode, DImode, "__avr32_s64_to_f64"); -+ set_conv_libfunc (sfloat_optab, SFmode, SImode, "__avr32_s32_to_f32"); -+ set_conv_libfunc (sfloat_optab, SFmode, DImode, "__avr32_s64_to_f32"); -+ set_conv_libfunc (ufloat_optab, DFmode, SImode, "__avr32_u32_to_f64"); -+ set_conv_libfunc (ufloat_optab, SFmode, SImode, "__avr32_u32_to_f32"); -+ /* TODO: Add these to gcc library functions */ -+ //set_conv_libfunc (ufloat_optab, DFmode, DImode, NULL); -+ //set_conv_libfunc (ufloat_optab, SFmode, DImode, NULL); -+ -+ /* Long long. Table 9. */ -+ set_optab_libfunc (smul_optab, DImode, "__avr32_mul64"); -+ set_optab_libfunc (sdiv_optab, DImode, "__avr32_sdiv64"); -+ set_optab_libfunc (udiv_optab, DImode, "__avr32_udiv64"); -+ set_optab_libfunc (smod_optab, DImode, "__avr32_smod64"); -+ set_optab_libfunc (umod_optab, DImode, "__avr32_umod64"); -+ set_optab_libfunc (ashl_optab, DImode, "__avr32_lsl64"); -+ set_optab_libfunc (lshr_optab, DImode, "__avr32_lsr64"); -+ set_optab_libfunc (ashr_optab, DImode, "__avr32_asr64"); -+ -+ /* Floating point library functions which have fast versions. */ -+ if ( TARGET_FAST_FLOAT ) -+ { -+ set_optab_libfunc (add_optab, DFmode, "__avr32_f64_add_fast"); -+ set_optab_libfunc (sub_optab, DFmode, "__avr32_f64_sub_fast"); -+ set_optab_libfunc (add_optab, SFmode, "__avr32_f32_add_fast"); -+ set_optab_libfunc (sub_optab, SFmode, "__avr32_f32_sub_fast"); -+ set_optab_libfunc (sdiv_optab, SFmode, "__avr32_f32_div_fast"); -+ } -+ else -+ { -+ set_optab_libfunc (add_optab, DFmode, "__avr32_f64_add"); -+ set_optab_libfunc (sub_optab, DFmode, "__avr32_f64_sub"); -+ set_optab_libfunc (add_optab, SFmode, "__avr32_f32_add"); -+ set_optab_libfunc (sub_optab, SFmode, "__avr32_f32_sub"); -+ set_optab_libfunc (sdiv_optab, SFmode, "__avr32_f32_div"); -+ } -+} -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-elf.h ---- gcc-4.2.1/gcc/config/avr32/avr32-elf.h 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-elf.h 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,84 @@ -+/* -+ Elf specific definitions. -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+ -+/***************************************************************************** -+ * Controlling the Compilator Driver, 'gcc' -+ *****************************************************************************/ -+ -+/* Run-time Target Specification. */ -+#undef TARGET_VERSION -+#define TARGET_VERSION fputs (" (AVR32 GNU with ELF)", stderr); -+ -+/* -+Another C string constant used much like LINK_SPEC. The -+difference between the two is that STARTFILE_SPEC is used at -+the very beginning of the command given to the linker. -+ -+If this macro is not defined, a default is provided that loads the -+standard C startup file from the usual place. See gcc.c. -+*/ -+#undef STARTFILE_SPEC -+#define STARTFILE_SPEC "crt0%O%s crti%O%s crtbegin%O%s" -+ -+#undef LINK_SPEC -+#define LINK_SPEC "%{muse-oscall:--defsym __do_not_use_oscall_coproc__=0} %{mrelax|O*:%{mno-relax|O0|O1: ;:--relax}} %{mpart=*:-mavr32elf_%*} %{mcpu=*:-mavr32elf_%*}" -+ -+ -+/* -+Another C string constant used much like LINK_SPEC. The -+difference between the two is that ENDFILE_SPEC is used at -+the very end of the command given to the linker. -+ -+Do not define this macro if it does not need to do anything. -+*/ -+#undef ENDFILE_SPEC -+#define ENDFILE_SPEC "crtend%O%s crtn%O%s" -+ -+ -+/* Target CPU builtins. */ -+#define TARGET_CPU_CPP_BUILTINS() \ -+ do \ -+ { \ -+ builtin_define ("__avr32__"); \ -+ builtin_define ("__AVR32__"); \ -+ builtin_define ("__AVR32_ELF__"); \ -+ builtin_define (avr32_part->macro); \ -+ builtin_define (avr32_arch->macro); \ -+ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) \ -+ builtin_define ("__AVR32_AVR32A__"); \ -+ else \ -+ builtin_define ("__AVR32_AVR32B__"); \ -+ if (TARGET_UNALIGNED_WORD) \ -+ builtin_define ("__AVR32_HAS_UNALIGNED_WORD__"); \ -+ if (TARGET_SIMD) \ -+ builtin_define ("__AVR32_HAS_SIMD__"); \ -+ if (TARGET_DSP) \ -+ builtin_define ("__AVR32_HAS_DSP__"); \ -+ if (TARGET_RMW) \ -+ builtin_define ("__AVR32_HAS_RMW__"); \ -+ if (TARGET_BRANCH_PRED) \ -+ builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ -+ if (TARGET_FAST_FLOAT) \ -+ builtin_define ("__AVR32_FAST_FLOAT__"); \ -+ } \ -+ while (0) -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.h ---- gcc-4.2.1/gcc/config/avr32/avr32.h 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.h 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,3281 @@ -+/* -+ Definitions of target machine for AVR32. -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ Initial porting by Anders �dland. -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+#ifndef GCC_AVR32_H -+#define GCC_AVR32_H -+ -+ -+#ifndef OBJECT_FORMAT_ELF -+#error avr32.h included before elfos.h -+#endif -+ -+#ifndef LOCAL_LABEL_PREFIX -+#define LOCAL_LABEL_PREFIX "." -+#endif -+ -+#ifndef SUBTARGET_CPP_SPEC -+#define SUBTARGET_CPP_SPEC "-D__ELF__" -+#endif -+ -+ -+extern struct rtx_def *avr32_compare_op0; -+extern struct rtx_def *avr32_compare_op1; -+ -+ -+extern struct rtx_def *avr32_acc_cache; -+ -+/* cache instruction op5 codes */ -+#define AVR32_CACHE_INVALIDATE_ICACHE 1 -+ -+/* These bits describe the different types of function supported -+ by the AVR32 backend. They are exclusive. ie a function cannot be both a -+ normal function and an interworked function, for example. Knowing the -+ type of a function is important for determining its prologue and -+ epilogue sequences. -+ Note value 7 is currently unassigned. Also note that the interrupt -+ function types all have bit 2 set, so that they can be tested for easily. -+ Note that 0 is deliberately chosen for AVR32_FT_UNKNOWN so that when the -+ machine_function structure is initialized (to zero) func_type will -+ default to unknown. This will force the first use of avr32_current_func_type -+ to call avr32_compute_func_type. */ -+#define AVR32_FT_UNKNOWN 0 /* Type has not yet been determined. -+ */ -+#define AVR32_FT_NORMAL 1 /* Your normal, straightforward -+ function. */ -+#define AVR32_FT_ACALL 2 /* An acall function. */ -+#define AVR32_FT_EXCEPTION_HANDLER 3 /* A C++ exception handler. */ -+#define AVR32_FT_ISR_FULL 4 /* A fully shadowed interrupt mode. */ -+#define AVR32_FT_ISR_HALF 5 /* A half shadowed interrupt mode. */ -+#define AVR32_FT_ISR_NONE 6 /* No shadow registers. */ -+ -+#define AVR32_FT_TYPE_MASK ((1 << 3) - 1) -+ -+/* In addition functions can have several type modifiers, -+ outlined by these bit masks: */ -+#define AVR32_FT_INTERRUPT (1 << 2) /* Note overlap with FT_ISR -+ and above. */ -+#define AVR32_FT_NAKED (1 << 3) /* No prologue or epilogue. */ -+#define AVR32_FT_VOLATILE (1 << 4) /* Does not return. */ -+#define AVR32_FT_NESTED (1 << 5) /* Embedded inside another -+ func. */ -+ -+/* Some macros to test these flags. */ -+#define AVR32_FUNC_TYPE(t) (t & AVR32_FT_TYPE_MASK) -+#define IS_INTERRUPT(t) (t & AVR32_FT_INTERRUPT) -+#define IS_VOLATILE(t) (t & AVR32_FT_VOLATILE) -+#define IS_NAKED(t) (t & AVR32_FT_NAKED) -+#define IS_NESTED(t) (t & AVR32_FT_NESTED) -+ -+ -+typedef struct minipool_labels -+GTY ((chain_next ("%h.next"), chain_prev ("%h.prev"))) -+{ -+ rtx label; -+ struct minipool_labels *prev; -+ struct minipool_labels *next; -+} minipool_labels; -+ -+/* A C structure for machine-specific, per-function data. -+ This is added to the cfun structure. */ -+ -+typedef struct machine_function -+GTY (()) -+{ -+ /* Records the type of the current function. */ -+ unsigned long func_type; -+ /* List of minipool labels, use for checking if code label is valid in a -+ memory expression */ -+ minipool_labels *minipool_label_head; -+ minipool_labels *minipool_label_tail; -+} machine_function; -+ -+/* Initialize data used by insn expanders. This is called from insn_emit, -+ once for every function before code is generated. */ -+#define INIT_EXPANDERS avr32_init_expanders () -+ -+/****************************************************************************** -+ * SPECS -+ *****************************************************************************/ -+ -+#ifndef ASM_SPEC -+#define ASM_SPEC "%{fpic:--pic} %{mrelax|O*:%{mno-relax|O0|O1: ;:--linkrelax}} %{march=*:-march=%*} %{mpart=*:-mpart=%*}" -+#endif -+ -+#ifndef MULTILIB_DEFAULTS -+#define MULTILIB_DEFAULTS { "march=ap", "" } -+#endif -+ -+/****************************************************************************** -+ * Run-time Target Specification -+ *****************************************************************************/ -+#ifndef TARGET_VERSION -+#define TARGET_VERSION fprintf(stderr, " (AVR32, GNU assembler syntax)"); -+#endif -+ -+/* Part types. Keep this in sync with the order of avr32_part_types in avr32.c*/ -+enum part_type -+{ -+ PART_TYPE_AVR32_NONE, -+ PART_TYPE_AVR32_AP7000, -+ PART_TYPE_AVR32_AP7010, -+ PART_TYPE_AVR32_AP7020, -+ PART_TYPE_AVR32_UC3A0256, -+ PART_TYPE_AVR32_UC3A0512, -+ PART_TYPE_AVR32_UC3A1128, -+ PART_TYPE_AVR32_UC3A1256, -+ PART_TYPE_AVR32_UC3A1512, -+ PART_TYPE_AVR32_UC3B064, -+ PART_TYPE_AVR32_UC3B0128, -+ PART_TYPE_AVR32_UC3B0256, -+ PART_TYPE_AVR32_UC3B164, -+ PART_TYPE_AVR32_UC3B1128, -+ PART_TYPE_AVR32_UC3B1256 -+}; -+ -+/* Microarchitectures. */ -+enum microarchitecture_type -+{ -+ UARCH_TYPE_AVR32A, -+ UARCH_TYPE_AVR32B, -+ UARCH_TYPE_NONE -+}; -+ -+/* Architectures types which specifies the pipeline. -+ Keep this in sync with avr32_arch_types in avr32.c -+ and the pipeline attribute in avr32.md */ -+enum architecture_type -+{ -+ ARCH_TYPE_AVR32_AP, -+ ARCH_TYPE_AVR32_UC, -+ ARCH_TYPE_AVR32_NONE -+}; -+ -+/* Flag specifying if the cpu has support for DSP instructions.*/ -+#define FLAG_AVR32_HAS_DSP (1 << 0) -+/* Flag specifying if the cpu has support for Read-Modify-Write -+ instructions.*/ -+#define FLAG_AVR32_HAS_RMW (1 << 1) -+/* Flag specifying if the cpu has support for SIMD instructions. */ -+#define FLAG_AVR32_HAS_SIMD (1 << 2) -+/* Flag specifying if the cpu has support for unaligned memory word access. */ -+#define FLAG_AVR32_HAS_UNALIGNED_WORD (1 << 3) -+/* Flag specifying if the cpu has support for branch prediction. */ -+#define FLAG_AVR32_HAS_BRANCH_PRED (1 << 4) -+/* Flag specifying if the cpu has support for a return stack. */ -+#define FLAG_AVR32_HAS_RETURN_STACK (1 << 5) -+/* Flag specifying if the cpu has caches. */ -+#define FLAG_AVR32_HAS_CACHES (1 << 6) -+ -+/* Structure for holding information about different avr32 CPUs/parts */ -+struct part_type_s -+{ -+ const char *const name; -+ enum part_type part_type; -+ enum architecture_type arch_type; -+ /* Must lie outside user's namespace. NULL == no macro. */ -+ const char *const macro; -+}; -+ -+/* Structure for holding information about different avr32 pipeline -+ architectures. */ -+struct arch_type_s -+{ -+ const char *const name; -+ enum architecture_type arch_type; -+ enum microarchitecture_type uarch_type; -+ const unsigned long feature_flags; -+ /* Must lie outside user's namespace. NULL == no macro. */ -+ const char *const macro; -+}; -+ -+extern const struct part_type_s *avr32_part; -+extern const struct arch_type_s *avr32_arch; -+ -+#define TARGET_SIMD (avr32_arch->feature_flags & FLAG_AVR32_HAS_SIMD) -+#define TARGET_DSP (avr32_arch->feature_flags & FLAG_AVR32_HAS_DSP) -+#define TARGET_RMW (avr32_arch->feature_flags & FLAG_AVR32_HAS_RMW) -+#define TARGET_UNALIGNED_WORD (avr32_arch->feature_flags & FLAG_AVR32_HAS_UNALIGNED_WORD) -+#define TARGET_BRANCH_PRED (avr32_arch->feature_flags & FLAG_AVR32_HAS_BRANCH_PRED) -+#define TARGET_RETURN_STACK (avr32_arch->feature_flags & FLAG_AVR32_HAS_RETURN_STACK) -+#define TARGET_CACHES (avr32_arch->feature_flags & FLAG_AVR32_HAS_CACHES) -+ -+#define CAN_DEBUG_WITHOUT_FP -+ -+/****************************************************************************** -+ * Storage Layout -+ *****************************************************************************/ -+ -+/* -+Define this macro to have the value 1 if the most significant bit in a -+byte has the lowest number; otherwise define it to have the value zero. -+This means that bit-field instructions count from the most significant -+bit. If the machine has no bit-field instructions, then this must still -+be defined, but it doesn't matter which value it is defined to. This -+macro need not be a constant. -+ -+This macro does not affect the way structure fields are packed into -+bytes or words; that is controlled by BYTES_BIG_ENDIAN. -+*/ -+#define BITS_BIG_ENDIAN 0 -+ -+/* -+Define this macro to have the value 1 if the most significant byte in a -+word has the lowest number. This macro need not be a constant. -+*/ -+/* -+ Data is stored in an big-endian way. -+*/ -+#define BYTES_BIG_ENDIAN 1 -+ -+/* -+Define this macro to have the value 1 if, in a multiword object, the -+most significant word has the lowest number. This applies to both -+memory locations and registers; GCC fundamentally assumes that the -+order of words in memory is the same as the order in registers. This -+macro need not be a constant. -+*/ -+/* -+ Data is stored in an bin-endian way. -+*/ -+#define WORDS_BIG_ENDIAN 1 -+ -+/* -+Define this macro if WORDS_BIG_ENDIAN is not constant. This must be a -+constant value with the same meaning as WORDS_BIG_ENDIAN, which will be -+used only when compiling libgcc2.c. Typically the value will be set -+based on preprocessor defines. -+*/ -+#define LIBGCC2_WORDS_BIG_ENDIAN WORDS_BIG_ENDIAN -+ -+/* -+Define this macro to have the value 1 if DFmode, XFmode or -+TFmode floating point numbers are stored in memory with the word -+containing the sign bit at the lowest address; otherwise define it to -+have the value 0. This macro need not be a constant. -+ -+You need not define this macro if the ordering is the same as for -+multi-word integers. -+*/ -+/* #define FLOAT_WORDS_BIG_ENDIAN 1 */ -+ -+/* -+Define this macro to be the number of bits in an addressable storage -+unit (byte); normally 8. -+*/ -+#define BITS_PER_UNIT 8 -+ -+/* -+Number of bits in a word; normally 32. -+*/ -+#define BITS_PER_WORD 32 -+ -+/* -+Maximum number of bits in a word. If this is undefined, the default is -+BITS_PER_WORD. Otherwise, it is the constant value that is the -+largest value that BITS_PER_WORD can have at run-time. -+*/ -+/* MAX_BITS_PER_WORD not defined*/ -+ -+/* -+Number of storage units in a word; normally 4. -+*/ -+#define UNITS_PER_WORD 4 -+ -+/* -+Minimum number of units in a word. If this is undefined, the default is -+UNITS_PER_WORD. Otherwise, it is the constant value that is the -+smallest value that UNITS_PER_WORD can have at run-time. -+*/ -+/* MIN_UNITS_PER_WORD not defined */ -+ -+/* -+Width of a pointer, in bits. You must specify a value no wider than the -+width of Pmode. If it is not equal to the width of Pmode, -+you must define POINTERS_EXTEND_UNSIGNED. -+*/ -+#define POINTER_SIZE 32 -+ -+/* -+A C expression whose value is greater than zero if pointers that need to be -+extended from being POINTER_SIZE bits wide to Pmode are to -+be zero-extended and zero if they are to be sign-extended. If the value -+is less then zero then there must be an "ptr_extend" instruction that -+extends a pointer from POINTER_SIZE to Pmode. -+ -+You need not define this macro if the POINTER_SIZE is equal -+to the width of Pmode. -+*/ -+/* #define POINTERS_EXTEND_UNSIGNED */ -+ -+/* -+A Macro to update M and UNSIGNEDP when an object whose type -+is TYPE and which has the specified mode and signedness is to be -+stored in a register. This macro is only called when TYPE is a -+scalar type. -+ -+On most RISC machines, which only have operations that operate on a full -+register, define this macro to set M to word_mode if -+M is an integer mode narrower than BITS_PER_WORD. In most -+cases, only integer modes should be widened because wider-precision -+floating-point operations are usually more expensive than their narrower -+counterparts. -+ -+For most machines, the macro definition does not change UNSIGNEDP. -+However, some machines, have instructions that preferentially handle -+either signed or unsigned quantities of certain modes. For example, on -+the DEC Alpha, 32-bit loads from memory and 32-bit add instructions -+sign-extend the result to 64 bits. On such machines, set -+UNSIGNEDP according to which kind of extension is more efficient. -+ -+Do not define this macro if it would never modify M. -+*/ -+#define PROMOTE_MODE(M, UNSIGNEDP, TYPE) \ -+ { \ -+ if (GET_MODE_CLASS (M) == MODE_INT \ -+ && GET_MODE_SIZE (M) < 4) \ -+ { \ -+ if (M == QImode) \ -+ UNSIGNEDP = 1; \ -+ else if (M == SImode) \ -+ UNSIGNEDP = 0; \ -+ (M) = SImode; \ -+ } \ -+ } -+ -+#define PROMOTE_FUNCTION_MODE(M, UNSIGNEDP, TYPE) \ -+ { \ -+ if (GET_MODE_CLASS (M) == MODE_INT \ -+ && GET_MODE_SIZE (M) < 4) \ -+ { \ -+ (M) = SImode; \ -+ } \ -+ } -+ -+/* Define if operations between registers always perform the operation -+ on the full register even if a narrower mode is specified. */ -+#define WORD_REGISTER_OPERATIONS -+ -+/* Define if loading in MODE, an integral mode narrower than BITS_PER_WORD -+ will either zero-extend or sign-extend. The value of this macro should -+ be the code that says which one of the two operations is implicitly -+ done, UNKNOWN if not known. */ -+#define LOAD_EXTEND_OP(MODE) \ -+ (((MODE) == QImode) ? ZERO_EXTEND \ -+ : ((MODE) == HImode) ? SIGN_EXTEND : UNKNOWN) -+ -+ -+/* -+Define this macro if the promotion described by PROMOTE_MODE -+should only be performed for outgoing function arguments or -+function return values, as specified by PROMOTE_FUNCTION_ARGS -+and PROMOTE_FUNCTION_RETURN, respectively. -+*/ -+/* #define PROMOTE_FOR_CALL_ONLY */ -+ -+/* -+Normal alignment required for function parameters on the stack, in -+bits. All stack parameters receive at least this much alignment -+regardless of data type. On most machines, this is the same as the -+size of an integer. -+*/ -+#define PARM_BOUNDARY 32 -+ -+/* -+Define this macro to the minimum alignment enforced by hardware for the -+stack pointer on this machine. The definition is a C expression for the -+desired alignment (measured in bits). This value is used as a default -+if PREFERRED_STACK_BOUNDARY is not defined. On most machines, -+this should be the same as PARM_BOUNDARY. -+*/ -+#define STACK_BOUNDARY 32 -+ -+/* -+Define this macro if you wish to preserve a certain alignment for the -+stack pointer, greater than what the hardware enforces. The definition -+is a C expression for the desired alignment (measured in bits). This -+macro must evaluate to a value equal to or larger than -+STACK_BOUNDARY. -+*/ -+#define PREFERRED_STACK_BOUNDARY (TARGET_FORCE_DOUBLE_ALIGN ? 64 : 32 ) -+ -+/* -+Alignment required for a function entry point, in bits. -+*/ -+#define FUNCTION_BOUNDARY 16 -+ -+/* -+Biggest alignment that any data type can require on this machine, in bits. -+*/ -+#define BIGGEST_ALIGNMENT (TARGET_FORCE_DOUBLE_ALIGN ? 64 : 32 ) -+ -+/* -+If defined, the smallest alignment, in bits, that can be given to an -+object that can be referenced in one operation, without disturbing any -+nearby object. Normally, this is BITS_PER_UNIT, but may be larger -+on machines that don't have byte or half-word store operations. -+*/ -+#define MINIMUM_ATOMIC_ALIGNMENT BITS_PER_UNIT -+ -+ -+/* -+An integer expression for the size in bits of the largest integer machine mode that -+should actually be used. All integer machine modes of this size or smaller can be -+used for structures and unions with the appropriate sizes. If this macro is undefined, -+GET_MODE_BITSIZE (DImode) is assumed.*/ -+#define MAX_FIXED_MODE_SIZE GET_MODE_BITSIZE (DImode) -+ -+ -+/* -+If defined, a C expression to compute the alignment given to a constant -+that is being placed in memory. CONSTANT is the constant and -+BASIC_ALIGN is the alignment that the object would ordinarily -+have. The value of this macro is used instead of that alignment to -+align the object. -+ -+If this macro is not defined, then BASIC_ALIGN is used. -+ -+The typical use of this macro is to increase alignment for string -+constants to be word aligned so that strcpy calls that copy -+constants can be done inline. -+*/ -+#define CONSTANT_ALIGNMENT(CONSTANT, BASIC_ALIGN) \ -+ ((TREE_CODE(CONSTANT) == STRING_CST) ? BITS_PER_WORD : BASIC_ALIGN) -+ -+/* Try to align string to a word. */ -+#define DATA_ALIGNMENT(TYPE, ALIGN) \ -+ ({(TREE_CODE (TYPE) == ARRAY_TYPE \ -+ && TYPE_MODE (TREE_TYPE (TYPE)) == QImode \ -+ && (ALIGN) < BITS_PER_WORD ? BITS_PER_WORD : (ALIGN));}) -+ -+/* Try to align local store strings to a word. */ -+#define LOCAL_ALIGNMENT(TYPE, ALIGN) \ -+ ({(TREE_CODE (TYPE) == ARRAY_TYPE \ -+ && TYPE_MODE (TREE_TYPE (TYPE)) == QImode \ -+ && (ALIGN) < BITS_PER_WORD ? BITS_PER_WORD : (ALIGN));}) -+ -+/* -+Define this macro to be the value 1 if instructions will fail to work -+if given data not on the nominal alignment. If instructions will merely -+go slower in that case, define this macro as 0. -+*/ -+#define STRICT_ALIGNMENT 1 -+ -+/* -+Define this if you wish to imitate the way many other C compilers handle -+alignment of bit-fields and the structures that contain them. -+ -+The behavior is that the type written for a bit-field (int, -+short, or other integer type) imposes an alignment for the -+entire structure, as if the structure really did contain an ordinary -+field of that type. In addition, the bit-field is placed within the -+structure so that it would fit within such a field, not crossing a -+boundary for it. -+ -+Thus, on most machines, a bit-field whose type is written as int -+would not cross a four-byte boundary, and would force four-byte -+alignment for the whole structure. (The alignment used may not be four -+bytes; it is controlled by the other alignment parameters.) -+ -+If the macro is defined, its definition should be a C expression; -+a nonzero value for the expression enables this behavior. -+ -+Note that if this macro is not defined, or its value is zero, some -+bit-fields may cross more than one alignment boundary. The compiler can -+support such references if there are insv, extv, and -+extzv insns that can directly reference memory. -+ -+The other known way of making bit-fields work is to define -+STRUCTURE_SIZE_BOUNDARY as large as BIGGEST_ALIGNMENT. -+Then every structure can be accessed with fullwords. -+ -+Unless the machine has bit-field instructions or you define -+STRUCTURE_SIZE_BOUNDARY that way, you must define -+PCC_BITFIELD_TYPE_MATTERS to have a nonzero value. -+ -+If your aim is to make GCC use the same conventions for laying out -+bit-fields as are used by another compiler, here is how to investigate -+what the other compiler does. Compile and run this program: -+ -+struct foo1 -+{ -+ char x; -+ char :0; -+ char y; -+}; -+ -+struct foo2 -+{ -+ char x; -+ int :0; -+ char y; -+}; -+ -+main () -+{ -+ printf ("Size of foo1 is %d\n", -+ sizeof (struct foo1)); -+ printf ("Size of foo2 is %d\n", -+ sizeof (struct foo2)); -+ exit (0); -+} -+ -+If this prints 2 and 5, then the compiler's behavior is what you would -+get from PCC_BITFIELD_TYPE_MATTERS. -+*/ -+#define PCC_BITFIELD_TYPE_MATTERS 1 -+ -+ -+/****************************************************************************** -+ * Layout of Source Language Data Types -+ *****************************************************************************/ -+ -+/* -+A C expression for the size in bits of the type int on the -+target machine. If you don't define this, the default is one word. -+*/ -+#define INT_TYPE_SIZE 32 -+ -+/* -+A C expression for the size in bits of the type short on the -+target machine. If you don't define this, the default is half a word. (If -+this would be less than one storage unit, it is rounded up to one unit.) -+*/ -+#define SHORT_TYPE_SIZE 16 -+ -+/* -+A C expression for the size in bits of the type long on the -+target machine. If you don't define this, the default is one word. -+*/ -+#define LONG_TYPE_SIZE 32 -+ -+ -+/* -+A C expression for the size in bits of the type long long on the -+target machine. If you don't define this, the default is two -+words. If you want to support GNU Ada on your machine, the value of this -+macro must be at least 64. -+*/ -+#define LONG_LONG_TYPE_SIZE 64 -+ -+/* -+A C expression for the size in bits of the type char on the -+target machine. If you don't define this, the default is -+BITS_PER_UNIT. -+*/ -+#define CHAR_TYPE_SIZE 8 -+ -+ -+/* -+A C expression for the size in bits of the C++ type bool and -+C99 type _Bool on the target machine. If you don't define -+this, and you probably shouldn't, the default is CHAR_TYPE_SIZE. -+*/ -+#define BOOL_TYPE_SIZE 8 -+ -+ -+/* -+An expression whose value is 1 or 0, according to whether the type -+char should be signed or unsigned by default. The user can -+always override this default with the options -fsigned-char -+and -funsigned-char. -+*/ -+/* We are using unsigned char */ -+#define DEFAULT_SIGNED_CHAR 0 -+ -+ -+/* -+A C expression for a string describing the name of the data type to use -+for size values. The typedef name size_t is defined using the -+contents of the string. -+ -+The string can contain more than one keyword. If so, separate them with -+spaces, and write first any length keyword, then unsigned if -+appropriate, and finally int. The string must exactly match one -+of the data type names defined in the function -+init_decl_processing in the file c-decl.c. You may not -+omit int or change the order - that would cause the compiler to -+crash on startup. -+ -+If you don't define this macro, the default is "long unsigned int". -+*/ -+#define SIZE_TYPE "long unsigned int" -+ -+/* -+A C expression for a string describing the name of the data type to use -+for the result of subtracting two pointers. The typedef name -+ptrdiff_t is defined using the contents of the string. See -+SIZE_TYPE above for more information. -+ -+If you don't define this macro, the default is "long int". -+*/ -+#define PTRDIFF_TYPE "long int" -+ -+ -+/* -+A C expression for the size in bits of the data type for wide -+characters. This is used in cpp, which cannot make use of -+WCHAR_TYPE. -+*/ -+#define WCHAR_TYPE_SIZE 32 -+ -+ -+/* -+A C expression for a string describing the name of the data type to -+use for wide characters passed to printf and returned from -+getwc. The typedef name wint_t is defined using the -+contents of the string. See SIZE_TYPE above for more -+information. -+ -+If you don't define this macro, the default is "unsigned int". -+*/ -+#define WINT_TYPE "unsigned int" -+ -+/* -+A C expression for a string describing the name of the data type that -+can represent any value of any standard or extended signed integer type. -+The typedef name intmax_t is defined using the contents of the -+string. See SIZE_TYPE above for more information. -+ -+If you don't define this macro, the default is the first of -+"int", "long int", or "long long int" that has as -+much precision as long long int. -+*/ -+#define INTMAX_TYPE "long long int" -+ -+/* -+A C expression for a string describing the name of the data type that -+can represent any value of any standard or extended unsigned integer -+type. The typedef name uintmax_t is defined using the contents -+of the string. See SIZE_TYPE above for more information. -+ -+If you don't define this macro, the default is the first of -+"unsigned int", "long unsigned int", or "long long unsigned int" -+that has as much precision as long long unsigned int. -+*/ -+#define UINTMAX_TYPE "long long unsigned int" -+ -+ -+/****************************************************************************** -+ * Register Usage -+ *****************************************************************************/ -+ -+/* Convert from gcc internal register number to register number -+ used in assembly code */ -+#define ASM_REGNUM(reg) (LAST_REGNUM - (reg)) -+#define ASM_FP_REGNUM(reg) (LAST_FP_REGNUM - (reg)) -+ -+/* Convert between register number used in assembly to gcc -+ internal register number */ -+#define INTERNAL_REGNUM(reg) (LAST_REGNUM - (reg)) -+#define INTERNAL_FP_REGNUM(reg) (LAST_FP_REGNUM - (reg)) -+ -+/** Basic Characteristics of Registers **/ -+ -+/* -+Number of hardware registers known to the compiler. They receive -+numbers 0 through FIRST_PSEUDO_REGISTER-1; thus, the first -+pseudo register's number really is assigned the number -+FIRST_PSEUDO_REGISTER. -+*/ -+#define FIRST_PSEUDO_REGISTER (LAST_FP_REGNUM + 1) -+ -+#define FIRST_REGNUM 0 -+#define LAST_REGNUM 15 -+#define NUM_FP_REGS 16 -+#define FIRST_FP_REGNUM 16 -+#define LAST_FP_REGNUM (16+NUM_FP_REGS-1) -+ -+/* -+An initializer that says which registers are used for fixed purposes -+all throughout the compiled code and are therefore not available for -+general allocation. These would include the stack pointer, the frame -+pointer (except on machines where that can be used as a general -+register when no frame pointer is needed), the program counter on -+machines where that is considered one of the addressable registers, -+and any other numbered register with a standard use. -+ -+This information is expressed as a sequence of numbers, separated by -+commas and surrounded by braces. The nth number is 1 if -+register n is fixed, 0 otherwise. -+ -+The table initialized from this macro, and the table initialized by -+the following one, may be overridden at run time either automatically, -+by the actions of the macro CONDITIONAL_REGISTER_USAGE, or by -+the user with the command options -ffixed-[reg], -+-fcall-used-[reg] and -fcall-saved-[reg]. -+*/ -+ -+/* The internal gcc register numbers are reversed -+ compared to the real register numbers since -+ gcc expects data types stored over multiple -+ registers in the register file to be big endian -+ if the memory layout is big endian. But this -+ is not the case for avr32 so we fake a big -+ endian register file. */ -+ -+#define FIXED_REGISTERS { \ -+ 1, /* Program Counter */ \ -+ 0, /* Link Register */ \ -+ 1, /* Stack Pointer */ \ -+ 0, /* r12 */ \ -+ 0, /* r11 */ \ -+ 0, /* r10 */ \ -+ 0, /* r9 */ \ -+ 0, /* r8 */ \ -+ 0, /* r7 */ \ -+ 0, /* r6 */ \ -+ 0, /* r5 */ \ -+ 0, /* r4 */ \ -+ 0, /* r3 */ \ -+ 0, /* r2 */ \ -+ 0, /* r1 */ \ -+ 0, /* r0 */ \ -+ 0, /* f15 */ \ -+ 0, /* f14 */ \ -+ 0, /* f13 */ \ -+ 0, /* f12 */ \ -+ 0, /* f11 */ \ -+ 0, /* f10 */ \ -+ 0, /* f9 */ \ -+ 0, /* f8 */ \ -+ 0, /* f7 */ \ -+ 0, /* f6 */ \ -+ 0, /* f5 */ \ -+ 0, /* f4 */ \ -+ 0, /* f3 */ \ -+ 0, /* f2*/ \ -+ 0, /* f1 */ \ -+ 0 /* f0 */ \ -+} -+ -+/* -+Like FIXED_REGISTERS but has 1 for each register that is -+clobbered (in general) by function calls as well as for fixed -+registers. This macro therefore identifies the registers that are not -+available for general allocation of values that must live across -+function calls. -+ -+If a register has 0 in CALL_USED_REGISTERS, the compiler -+automatically saves it on function entry and restores it on function -+exit, if the register is used within the function. -+*/ -+#define CALL_USED_REGISTERS { \ -+ 1, /* Program Counter */ \ -+ 0, /* Link Register */ \ -+ 1, /* Stack Pointer */ \ -+ 1, /* r12 */ \ -+ 1, /* r11 */ \ -+ 1, /* r10 */ \ -+ 1, /* r9 */ \ -+ 1, /* r8 */ \ -+ 0, /* r7 */ \ -+ 0, /* r6 */ \ -+ 0, /* r5 */ \ -+ 0, /* r4 */ \ -+ 0, /* r3 */ \ -+ 0, /* r2 */ \ -+ 0, /* r1 */ \ -+ 0, /* r0 */ \ -+ 1, /* f15 */ \ -+ 1, /* f14 */ \ -+ 1, /* f13 */ \ -+ 1, /* f12 */ \ -+ 1, /* f11 */ \ -+ 1, /* f10 */ \ -+ 1, /* f9 */ \ -+ 1, /* f8 */ \ -+ 0, /* f7 */ \ -+ 0, /* f6 */ \ -+ 0, /* f5 */ \ -+ 0, /* f4 */ \ -+ 0, /* f3 */ \ -+ 0, /* f2*/ \ -+ 0, /* f1*/ \ -+ 0, /* f0 */ \ -+} -+ -+/* Interrupt functions can only use registers that have already been -+ saved by the prologue, even if they would normally be -+ call-clobbered. */ -+#define HARD_REGNO_RENAME_OK(SRC, DST) \ -+ (! IS_INTERRUPT (cfun->machine->func_type) || \ -+ regs_ever_live[DST]) -+ -+ -+/* -+Zero or more C statements that may conditionally modify five variables -+fixed_regs, call_used_regs, global_regs, -+reg_names, and reg_class_contents, to take into account -+any dependence of these register sets on target flags. The first three -+of these are of type char [] (interpreted as Boolean vectors). -+global_regs is a const char *[], and -+reg_class_contents is a HARD_REG_SET. Before the macro is -+called, fixed_regs, call_used_regs, -+reg_class_contents, and reg_names have been initialized -+from FIXED_REGISTERS, CALL_USED_REGISTERS, -+REG_CLASS_CONTENTS, and REGISTER_NAMES, respectively. -+global_regs has been cleared, and any -ffixed-[reg], -+-fcall-used-[reg] and -fcall-saved-[reg] -+command options have been applied. -+ -+You need not define this macro if it has no work to do. -+ -+If the usage of an entire class of registers depends on the target -+flags, you may indicate this to GCC by using this macro to modify -+fixed_regs and call_used_regs to 1 for each of the -+registers in the classes which should not be used by GCC. Also define -+the macro REG_CLASS_FROM_LETTER to return NO_REGS if it -+is called with a letter for a class that shouldn't be used. -+ -+ (However, if this class is not included in GENERAL_REGS and all -+of the insn patterns whose constraints permit this class are -+controlled by target switches, then GCC will automatically avoid using -+these registers when the target switches are opposed to them.) -+*/ -+#define CONDITIONAL_REGISTER_USAGE \ -+ do \ -+ { \ -+ int regno; \ -+ \ -+ if (TARGET_SOFT_FLOAT) \ -+ { \ -+ for (regno = FIRST_FP_REGNUM; \ -+ regno <= LAST_FP_REGNUM; ++regno) \ -+ fixed_regs[regno] = call_used_regs[regno] = 1; \ -+ } \ -+ if (flag_pic) \ -+ { \ -+ fixed_regs[PIC_OFFSET_TABLE_REGNUM] = 1; \ -+ call_used_regs[PIC_OFFSET_TABLE_REGNUM] = 1; \ -+ } \ -+ } \ -+ while (0) -+ -+ -+/* -+If the program counter has a register number, define this as that -+register number. Otherwise, do not define it. -+*/ -+ -+#define LAST_AVR32_REGNUM 16 -+ -+ -+/** Order of Allocation of Registers **/ -+ -+/* -+If defined, an initializer for a vector of integers, containing the -+numbers of hard registers in the order in which GCC should prefer -+to use them (from most preferred to least). -+ -+If this macro is not defined, registers are used lowest numbered first -+(all else being equal). -+ -+One use of this macro is on machines where the highest numbered -+registers must always be saved and the save-multiple-registers -+instruction supports only sequences of consecutive registers. On such -+machines, define REG_ALLOC_ORDER to be an initializer that lists -+the highest numbered allocable register first. -+*/ -+#define REG_ALLOC_ORDER \ -+{ \ -+ INTERNAL_REGNUM(8), \ -+ INTERNAL_REGNUM(9), \ -+ INTERNAL_REGNUM(10), \ -+ INTERNAL_REGNUM(11), \ -+ INTERNAL_REGNUM(12), \ -+ LR_REGNUM, \ -+ INTERNAL_REGNUM(7), \ -+ INTERNAL_REGNUM(6), \ -+ INTERNAL_REGNUM(5), \ -+ INTERNAL_REGNUM(4), \ -+ INTERNAL_REGNUM(3), \ -+ INTERNAL_REGNUM(2), \ -+ INTERNAL_REGNUM(1), \ -+ INTERNAL_REGNUM(0), \ -+ INTERNAL_FP_REGNUM(15), \ -+ INTERNAL_FP_REGNUM(14), \ -+ INTERNAL_FP_REGNUM(13), \ -+ INTERNAL_FP_REGNUM(12), \ -+ INTERNAL_FP_REGNUM(11), \ -+ INTERNAL_FP_REGNUM(10), \ -+ INTERNAL_FP_REGNUM(9), \ -+ INTERNAL_FP_REGNUM(8), \ -+ INTERNAL_FP_REGNUM(7), \ -+ INTERNAL_FP_REGNUM(6), \ -+ INTERNAL_FP_REGNUM(5), \ -+ INTERNAL_FP_REGNUM(4), \ -+ INTERNAL_FP_REGNUM(3), \ -+ INTERNAL_FP_REGNUM(2), \ -+ INTERNAL_FP_REGNUM(1), \ -+ INTERNAL_FP_REGNUM(0), \ -+ SP_REGNUM, \ -+ PC_REGNUM \ -+} -+ -+ -+/** How Values Fit in Registers **/ -+ -+/* -+A C expression for the number of consecutive hard registers, starting -+at register number REGNO, required to hold a value of mode -+MODE. -+ -+On a machine where all registers are exactly one word, a suitable -+definition of this macro is -+ -+#define HARD_REGNO_NREGS(REGNO, MODE) \ -+ ((GET_MODE_SIZE (MODE) + UNITS_PER_WORD - 1) \ -+ / UNITS_PER_WORD) -+*/ -+#define HARD_REGNO_NREGS(REGNO, MODE) \ -+ ((unsigned int)((GET_MODE_SIZE(MODE) + UNITS_PER_WORD -1 ) / UNITS_PER_WORD)) -+ -+/* -+A C expression that is nonzero if it is permissible to store a value -+of mode MODE in hard register number REGNO (or in several -+registers starting with that one). For a machine where all registers -+are equivalent, a suitable definition is -+ -+ #define HARD_REGNO_MODE_OK(REGNO, MODE) 1 -+ -+You need not include code to check for the numbers of fixed registers, -+because the allocation mechanism considers them to be always occupied. -+ -+On some machines, double-precision values must be kept in even/odd -+register pairs. You can implement that by defining this macro to reject -+odd register numbers for such modes. -+ -+The minimum requirement for a mode to be OK in a register is that the -+mov[mode] instruction pattern support moves between the -+register and other hard register in the same class and that moving a -+value into the register and back out not alter it. -+ -+Since the same instruction used to move word_mode will work for -+all narrower integer modes, it is not necessary on any machine for -+HARD_REGNO_MODE_OK to distinguish between these modes, provided -+you define patterns movhi, etc., to take advantage of this. This -+is useful because of the interaction between HARD_REGNO_MODE_OK -+and MODES_TIEABLE_P; it is very desirable for all integer modes -+to be tieable. -+ -+Many machines have special registers for floating point arithmetic. -+Often people assume that floating point machine modes are allowed only -+in floating point registers. This is not true. Any registers that -+can hold integers can safely hold a floating point machine -+mode, whether or not floating arithmetic can be done on it in those -+registers. Integer move instructions can be used to move the values. -+ -+On some machines, though, the converse is true: fixed-point machine -+modes may not go in floating registers. This is true if the floating -+registers normalize any value stored in them, because storing a -+non-floating value there would garble it. In this case, -+HARD_REGNO_MODE_OK should reject fixed-point machine modes in -+floating registers. But if the floating registers do not automatically -+normalize, if you can store any bit pattern in one and retrieve it -+unchanged without a trap, then any machine mode may go in a floating -+register, so you can define this macro to say so. -+ -+The primary significance of special floating registers is rather that -+they are the registers acceptable in floating point arithmetic -+instructions. However, this is of no concern to -+HARD_REGNO_MODE_OK. You handle it by writing the proper -+constraints for those instructions. -+ -+On some machines, the floating registers are especially slow to access, -+so that it is better to store a value in a stack frame than in such a -+register if floating point arithmetic is not being done. As long as the -+floating registers are not in class GENERAL_REGS, they will not -+be used unless some pattern's constraint asks for one. -+*/ -+#define HARD_REGNO_MODE_OK(REGNO, MODE) avr32_hard_regno_mode_ok(REGNO, MODE) -+ -+/* -+A C expression that is nonzero if a value of mode -+MODE1 is accessible in mode MODE2 without copying. -+ -+If HARD_REGNO_MODE_OK(R, MODE1) and -+HARD_REGNO_MODE_OK(R, MODE2) are always the same for -+any R, then MODES_TIEABLE_P(MODE1, MODE2) -+should be nonzero. If they differ for any R, you should define -+this macro to return zero unless some other mechanism ensures the -+accessibility of the value in a narrower mode. -+ -+You should define this macro to return nonzero in as many cases as -+possible since doing so will allow GCC to perform better register -+allocation. -+*/ -+#define MODES_TIEABLE_P(MODE1, MODE2) \ -+ (GET_MODE_CLASS (MODE1) == GET_MODE_CLASS (MODE2)) -+ -+ -+ -+/****************************************************************************** -+ * Register Classes -+ *****************************************************************************/ -+ -+/* -+An enumeral type that must be defined with all the register class names -+as enumeral values. NO_REGS must be first. ALL_REGS -+must be the last register class, followed by one more enumeral value, -+LIM_REG_CLASSES, which is not a register class but rather -+tells how many classes there are. -+ -+Each register class has a number, which is the value of casting -+the class name to type int. The number serves as an index -+in many of the tables described below. -+*/ -+enum reg_class -+{ -+ NO_REGS, -+ GENERAL_REGS, -+ FP_REGS, -+ ALL_REGS, -+ LIM_REG_CLASSES -+}; -+ -+/* -+The number of distinct register classes, defined as follows: -+ #define N_REG_CLASSES (int) LIM_REG_CLASSES -+*/ -+#define N_REG_CLASSES (int)LIM_REG_CLASSES -+ -+/* -+An initializer containing the names of the register classes as C string -+constants. These names are used in writing some of the debugging dumps. -+*/ -+#define REG_CLASS_NAMES \ -+{ \ -+ "NO_REGS", \ -+ "GENERAL_REGS", \ -+ "FLOATING_POINT_REGS", \ -+ "ALL_REGS" \ -+} -+ -+/* -+An initializer containing the contents of the register classes, as integers -+which are bit masks. The nth integer specifies the contents of class -+n. The way the integer mask is interpreted is that -+register r is in the class if mask & (1 << r) is 1. -+ -+When the machine has more than 32 registers, an integer does not suffice. -+Then the integers are replaced by sub-initializers, braced groupings containing -+several integers. Each sub-initializer must be suitable as an initializer -+for the type HARD_REG_SET which is defined in hard-reg-set.h. -+In this situation, the first integer in each sub-initializer corresponds to -+registers 0 through 31, the second integer to registers 32 through 63, and -+so on. -+*/ -+#define REG_CLASS_CONTENTS { \ -+ {0x00000000}, /* NO_REGS */ \ -+ {0x0000FFFF}, /* GENERAL_REGS */ \ -+ {0xFFFF0000}, /* FP_REGS */ \ -+ {0x7FFFFFFF}, /* ALL_REGS */ \ -+} -+ -+ -+/* -+A C expression whose value is a register class containing hard register -+REGNO. In general there is more than one such class; choose a class -+which is minimal, meaning that no smaller class also contains the -+register. -+*/ -+#define REGNO_REG_CLASS(REGNO) ((REGNO < 16) ? GENERAL_REGS : FP_REGS) -+ -+/* -+A macro whose definition is the name of the class to which a valid -+base register must belong. A base register is one used in an address -+which is the register value plus a displacement. -+*/ -+#define BASE_REG_CLASS GENERAL_REGS -+ -+/* -+This is a variation of the BASE_REG_CLASS macro which allows -+the selection of a base register in a mode depenedent manner. If -+mode is VOIDmode then it should return the same value as -+BASE_REG_CLASS. -+*/ -+#define MODE_BASE_REG_CLASS(MODE) BASE_REG_CLASS -+ -+/* -+A macro whose definition is the name of the class to which a valid -+index register must belong. An index register is one used in an -+address where its value is either multiplied by a scale factor or -+added to another register (as well as added to a displacement). -+*/ -+#define INDEX_REG_CLASS BASE_REG_CLASS -+ -+/* -+A C expression which defines the machine-dependent operand constraint -+letters for register classes. If CHAR is such a letter, the -+value should be the register class corresponding to it. Otherwise, -+the value should be NO_REGS. The register letter r, -+corresponding to class GENERAL_REGS, will not be passed -+to this macro; you do not need to handle it. -+*/ -+#define REG_CLASS_FROM_LETTER(CHAR) ((CHAR) == 'f' ? FP_REGS : NO_REGS) -+ -+ -+/* These assume that REGNO is a hard or pseudo reg number. -+ They give nonzero only if REGNO is a hard reg of the suitable class -+ or a pseudo reg currently allocated to a suitable hard reg. -+ Since they use reg_renumber, they are safe only once reg_renumber -+ has been allocated, which happens in local-alloc.c. */ -+#define TEST_REGNO(R, TEST, VALUE) \ -+ ((R TEST VALUE) || ((unsigned) reg_renumber[R] TEST VALUE)) -+ -+/* -+A C expression which is nonzero if register number num is suitable for use as a base -+register in operand addresses. It may be either a suitable hard register or a pseudo -+register that has been allocated such a hard register. -+*/ -+#define REGNO_OK_FOR_BASE_P(NUM) TEST_REGNO(NUM, <=, LAST_REGNUM) -+ -+/* -+A C expression which is nonzero if register number NUM is -+suitable for use as an index register in operand addresses. It may be -+either a suitable hard register or a pseudo register that has been -+allocated such a hard register. -+ -+The difference between an index register and a base register is that -+the index register may be scaled. If an address involves the sum of -+two registers, neither one of them scaled, then either one may be -+labeled the ``base'' and the other the ``index''; but whichever -+labeling is used must fit the machine's constraints of which registers -+may serve in each capacity. The compiler will try both labelings, -+looking for one that is valid, and will reload one or both registers -+only if neither labeling works. -+*/ -+#define REGNO_OK_FOR_INDEX_P(NUM) TEST_REGNO(NUM, <=, LAST_REGNUM) -+ -+/* -+A C expression that places additional restrictions on the register class -+to use when it is necessary to copy value X into a register in class -+CLASS. The value is a register class; perhaps CLASS, or perhaps -+another, smaller class. On many machines, the following definition is -+safe: #define PREFERRED_RELOAD_CLASS(X,CLASS) CLASS -+ -+Sometimes returning a more restrictive class makes better code. For -+example, on the 68000, when X is an integer constant that is in range -+for a 'moveq' instruction, the value of this macro is always -+DATA_REGS as long as CLASS includes the data registers. -+Requiring a data register guarantees that a 'moveq' will be used. -+ -+If X is a const_double, by returning NO_REGS -+you can force X into a memory constant. This is useful on -+certain machines where immediate floating values cannot be loaded into -+certain kinds of registers. -+*/ -+#define PREFERRED_RELOAD_CLASS(X, CLASS) CLASS -+ -+ -+ -+/* -+A C expression for the maximum number of consecutive registers -+of class CLASS needed to hold a value of mode MODE. -+ -+This is closely related to the macro HARD_REGNO_NREGS. In fact, -+the value of the macro CLASS_MAX_NREGS(CLASS, MODE) -+should be the maximum value of HARD_REGNO_NREGS(REGNO, MODE) -+for all REGNO values in the class CLASS. -+ -+This macro helps control the handling of multiple-word values -+in the reload pass. -+*/ -+#define CLASS_MAX_NREGS(CLASS, MODE) /* ToDo:fixme */ \ -+ (unsigned int)((GET_MODE_SIZE(MODE) + UNITS_PER_WORD - 1) / UNITS_PER_WORD) -+ -+ -+/* -+ Using CONST_OK_FOR_CONSTRAINT_P instead of CONS_OK_FOR_LETTER_P -+ in order to support constraints with more than one letter. -+ Only two letters are then used for constant constraints, -+ the letter 'K' and the letter 'I'. The constraint starting with -+ these letters must consist of four characters. The character following -+ 'K' or 'I' must be either 'u' (unsigned) or 's' (signed) to specify -+ if the constant is zero or sign extended. The last two characters specify -+ the length in bits of the constant. The base constraint letter 'I' means -+ that this is an negated constant, meaning that actually -VAL should be -+ checked to lie withing the valid range instead of VAL which is used when -+ 'K' is the base constraint letter. -+ -+*/ -+ -+#define CONSTRAINT_LEN(C, STR) \ -+ ( ((C) == 'K' || (C) == 'I') ? 4 : \ -+ ((C) == 'R') ? 5 : \ -+ ((C) == 'N' || (C) == 'O' || \ -+ (C) == 'P' || (C) == 'L' || (C) == 'J') ? -1 : \ -+ DEFAULT_CONSTRAINT_LEN((C), (STR)) ) -+ -+#define CONST_OK_FOR_CONSTRAINT_P(VALUE, C, STR) \ -+ avr32_const_ok_for_constraint_p(VALUE, C, STR) -+ -+/* -+A C expression that defines the machine-dependent operand constraint -+letters that specify particular ranges of const_double values ('G' or 'H'). -+ -+If C is one of those letters, the expression should check that -+VALUE, an RTX of code const_double, is in the appropriate -+range and return 1 if so, 0 otherwise. If C is not one of those -+letters, the value should be 0 regardless of VALUE. -+ -+const_double is used for all floating-point constants and for -+DImode fixed-point constants. A given letter can accept either -+or both kinds of values. It can use GET_MODE to distinguish -+between these kinds. -+*/ -+#define CONST_DOUBLE_OK_FOR_LETTER_P(OP, C) \ -+ ((C) == 'G' ? avr32_const_double_immediate(OP) : 0) -+ -+/* -+A C expression that defines the optional machine-dependent constraint -+letters that can be used to segregate specific types of operands, usually -+memory references, for the target machine. Any letter that is not -+elsewhere defined and not matched by REG_CLASS_FROM_LETTER -+may be used. Normally this macro will not be defined. -+ -+If it is required for a particular target machine, it should return 1 -+if VALUE corresponds to the operand type represented by the -+constraint letter C. If C is not defined as an extra -+constraint, the value returned should be 0 regardless of VALUE. -+ -+For example, on the ROMP, load instructions cannot have their output -+in r0 if the memory reference contains a symbolic address. Constraint -+letter 'Q' is defined as representing a memory address that does -+not contain a symbolic address. An alternative is specified with -+a 'Q' constraint on the input and 'r' on the output. The next -+alternative specifies 'm' on the input and a register class that -+does not include r0 on the output. -+*/ -+#define EXTRA_CONSTRAINT_STR(OP, C, STR) \ -+ ((C) == 'W' ? avr32_address_operand(OP, GET_MODE(OP)) : \ -+ (C) == 'R' ? (avr32_indirect_register_operand(OP, GET_MODE(OP)) || \ -+ (avr32_imm_disp_memory_operand(OP, GET_MODE(OP)) \ -+ && avr32_const_ok_for_constraint_p( \ -+ INTVAL(XEXP(XEXP(OP, 0), 1)), \ -+ (STR)[1], &(STR)[1]))) : \ -+ (C) == 'S' ? avr32_indexed_memory_operand(OP, GET_MODE(OP)) : \ -+ (C) == 'T' ? avr32_const_pool_ref_operand(OP, GET_MODE(OP)) : \ -+ (C) == 'U' ? SYMBOL_REF_RCALL_FUNCTION_P(OP) : \ -+ (C) == 'Z' ? avr32_cop_memory_operand(OP, GET_MODE(OP)) : \ -+ 0) -+ -+ -+#define EXTRA_MEMORY_CONSTRAINT(C, STR) ( ((C) == 'R') || \ -+ ((C) == 'S') || \ -+ ((C) == 'Z') ) -+ -+ -+/* Returns nonzero if op is a function SYMBOL_REF which -+ can be called using an rcall instruction */ -+#define SYMBOL_REF_RCALL_FUNCTION_P(op) \ -+ ( GET_CODE(op) == SYMBOL_REF \ -+ && SYMBOL_REF_FUNCTION_P(op) \ -+ && SYMBOL_REF_LOCAL_P(op) \ -+ && !SYMBOL_REF_EXTERNAL_P(op) \ -+ && !TARGET_HAS_ASM_ADDR_PSEUDOS ) -+ -+/****************************************************************************** -+ * Stack Layout and Calling Conventions -+ *****************************************************************************/ -+ -+/** Basic Stack Layout **/ -+ -+/* -+Define this macro if pushing a word onto the stack moves the stack -+pointer to a smaller address. -+ -+When we say, ``define this macro if ...,'' it means that the -+compiler checks this macro only with #ifdef so the precise -+definition used does not matter. -+*/ -+/* pushm decrece SP: *(--SP) <-- Rx */ -+#define STACK_GROWS_DOWNWARD -+ -+/* -+This macro defines the operation used when something is pushed -+on the stack. In RTL, a push operation will be -+(set (mem (STACK_PUSH_CODE (reg sp))) ...) -+ -+The choices are PRE_DEC, POST_DEC, PRE_INC, -+and POST_INC. Which of these is correct depends on -+the stack direction and on whether the stack pointer points -+to the last item on the stack or whether it points to the -+space for the next item on the stack. -+ -+The default is PRE_DEC when STACK_GROWS_DOWNWARD is -+defined, which is almost always right, and PRE_INC otherwise, -+which is often wrong. -+*/ -+/* pushm: *(--SP) <-- Rx */ -+#define STACK_PUSH_CODE PRE_DEC -+ -+/* Define this to nonzero if the nominal address of the stack frame -+ is at the high-address end of the local variables; -+ that is, each additional local variable allocated -+ goes at a more negative offset in the frame. */ -+#define FRAME_GROWS_DOWNWARD 1 -+ -+ -+/* -+Offset from the frame pointer to the first local variable slot to be allocated. -+ -+If FRAME_GROWS_DOWNWARD, find the next slot's offset by -+subtracting the first slot's length from STARTING_FRAME_OFFSET. -+Otherwise, it is found by adding the length of the first slot to the -+value STARTING_FRAME_OFFSET. -+ (i'm not sure if the above is still correct.. had to change it to get -+ rid of an overfull. --mew 2feb93 ) -+*/ -+#define STARTING_FRAME_OFFSET 0 -+ -+/* -+Offset from the stack pointer register to the first location at which -+outgoing arguments are placed. If not specified, the default value of -+zero is used. This is the proper value for most machines. -+ -+If ARGS_GROW_DOWNWARD, this is the offset to the location above -+the first location at which outgoing arguments are placed. -+*/ -+#define STACK_POINTER_OFFSET 0 -+ -+/* -+Offset from the argument pointer register to the first argument's -+address. On some machines it may depend on the data type of the -+function. -+ -+If ARGS_GROW_DOWNWARD, this is the offset to the location above -+the first argument's address. -+*/ -+#define FIRST_PARM_OFFSET(FUNDECL) 0 -+ -+ -+/* -+A C expression whose value is RTL representing the address in a stack -+frame where the pointer to the caller's frame is stored. Assume that -+FRAMEADDR is an RTL expression for the address of the stack frame -+itself. -+ -+If you don't define this macro, the default is to return the value -+of FRAMEADDR - that is, the stack frame address is also the -+address of the stack word that points to the previous frame. -+*/ -+#define DYNAMIC_CHAIN_ADDRESS(FRAMEADDR) plus_constant ((FRAMEADDR), 4) -+ -+ -+/* -+A C expression whose value is RTL representing the value of the return -+address for the frame COUNT steps up from the current frame, after -+the prologue. FRAMEADDR is the frame pointer of the COUNT -+frame, or the frame pointer of the COUNT - 1 frame if -+RETURN_ADDR_IN_PREVIOUS_FRAME is defined. -+ -+The value of the expression must always be the correct address when -+COUNT is zero, but may be NULL_RTX if there is not way to -+determine the return address of other frames. -+*/ -+#define RETURN_ADDR_RTX(COUNT, FRAMEADDR) avr32_return_addr(COUNT, FRAMEADDR) -+ -+ -+/* -+A C expression whose value is RTL representing the location of the -+incoming return address at the beginning of any function, before the -+prologue. This RTL is either a REG, indicating that the return -+value is saved in 'REG', or a MEM representing a location in -+the stack. -+ -+You only need to define this macro if you want to support call frame -+debugging information like that provided by DWARF 2. -+ -+If this RTL is a REG, you should also define -+DWARF_FRAME_RETURN_COLUMN to DWARF_FRAME_REGNUM (REGNO). -+*/ -+#define INCOMING_RETURN_ADDR_RTX gen_rtx_REG (Pmode, LR_REGNUM) -+ -+ -+ -+/* -+A C expression whose value is an integer giving the offset, in bytes, -+from the value of the stack pointer register to the top of the stack -+frame at the beginning of any function, before the prologue. The top of -+the frame is defined to be the value of the stack pointer in the -+previous frame, just before the call instruction. -+ -+You only need to define this macro if you want to support call frame -+debugging information like that provided by DWARF 2. -+*/ -+#define INCOMING_FRAME_SP_OFFSET 0 -+ -+ -+/** Exception Handling Support **/ -+ -+/* Use setjump/longjump for exception handling. */ -+#define DWARF2_UNWIND_INFO 0 -+#define MUST_USE_SJLJ_EXCEPTIONS 1 -+ -+/* -+A C expression whose value is the Nth register number used for -+data by exception handlers, or INVALID_REGNUM if fewer than -+N registers are usable. -+ -+The exception handling library routines communicate with the exception -+handlers via a set of agreed upon registers. Ideally these registers -+should be call-clobbered; it is possible to use call-saved registers, -+but may negatively impact code size. The target must support at least -+2 data registers, but should define 4 if there are enough free registers. -+ -+You must define this macro if you want to support call frame exception -+handling like that provided by DWARF 2. -+*/ -+/* -+ Use r9-r11 -+*/ -+#define EH_RETURN_DATA_REGNO(N) \ -+ ((N<3) ? INTERNAL_REGNUM(N+9) : INVALID_REGNUM) -+ -+/* -+A C expression whose value is RTL representing a location in which -+to store a stack adjustment to be applied before function return. -+This is used to unwind the stack to an exception handler's call frame. -+It will be assigned zero on code paths that return normally. -+ -+Typically this is a call-clobbered hard register that is otherwise -+untouched by the epilogue, but could also be a stack slot. -+ -+You must define this macro if you want to support call frame exception -+handling like that provided by DWARF 2. -+*/ -+/* -+ Use r8 -+*/ -+#define EH_RETURN_STACKADJ_REGNO INTERNAL_REGNUM(8) -+#define EH_RETURN_STACKADJ_RTX gen_rtx_REG(SImode, EH_RETURN_STACKADJ_REGNO) -+ -+/* -+A C expression whose value is RTL representing a location in which -+to store the address of an exception handler to which we should -+return. It will not be assigned on code paths that return normally. -+ -+Typically this is the location in the call frame at which the normal -+return address is stored. For targets that return by popping an -+address off the stack, this might be a memory address just below -+the target call frame rather than inside the current call -+frame. EH_RETURN_STACKADJ_RTX will have already been assigned, -+so it may be used to calculate the location of the target call frame. -+ -+Some targets have more complex requirements than storing to an -+address calculable during initial code generation. In that case -+the eh_return instruction pattern should be used instead. -+ -+If you want to support call frame exception handling, you must -+define either this macro or the eh_return instruction pattern. -+*/ -+/* -+ We define the eh_return instruction pattern, so this isn't needed. -+*/ -+/* #define EH_RETURN_HANDLER_RTX gen_rtx_REG(Pmode, RET_REGISTER) */ -+ -+/* -+ This macro chooses the encoding of pointers embedded in the -+ exception handling sections. If at all possible, this should be -+ defined such that the exception handling section will not require -+ dynamic relocations, and so may be read-only. -+ -+ code is 0 for data, 1 for code labels, 2 for function -+ pointers. global is true if the symbol may be affected by dynamic -+ relocations. The macro should return a combination of the DW_EH_PE_* -+ defines as found in dwarf2.h. -+ -+ If this macro is not defined, pointers will not be encoded but -+ represented directly. -+*/ -+#define ASM_PREFERRED_EH_DATA_FORMAT(CODE, GLOBAL) \ -+ ((flag_pic && (GLOBAL) ? DW_EH_PE_indirect : 0) \ -+ | (flag_pic ? DW_EH_PE_pcrel : DW_EH_PE_absptr) \ -+ | DW_EH_PE_sdata4) -+ -+/* ToDo: The rest of this subsection */ -+ -+/** Specifying How Stack Checking is Done **/ -+/* ToDo: All in this subsection */ -+ -+/** Registers That Address the Stack Frame **/ -+ -+/* -+The register number of the stack pointer register, which must also be a -+fixed register according to FIXED_REGISTERS. On most machines, -+the hardware determines which register this is. -+*/ -+/* Using r13 as stack pointer. */ -+#define STACK_POINTER_REGNUM INTERNAL_REGNUM(13) -+ -+/* -+The register number of the frame pointer register, which is used to -+access automatic variables in the stack frame. On some machines, the -+hardware determines which register this is. On other machines, you can -+choose any register you wish for this purpose. -+*/ -+/* Use r7 */ -+#define FRAME_POINTER_REGNUM INTERNAL_REGNUM(7) -+ -+ -+ -+/* -+The register number of the arg pointer register, which is used to access -+the function's argument list. On some machines, this is the same as the -+frame pointer register. On some machines, the hardware determines which -+register this is. On other machines, you can choose any register you -+wish for this purpose. If this is not the same register as the frame -+pointer register, then you must mark it as a fixed register according to -+FIXED_REGISTERS, or arrange to be able to eliminate it (see Section -+10.10.5 [Elimination], page 224). -+*/ -+/* Using r5 */ -+#define ARG_POINTER_REGNUM INTERNAL_REGNUM(4) -+ -+ -+/* -+Register numbers used for passing a function's static chain pointer. If -+register windows are used, the register number as seen by the called -+function is STATIC_CHAIN_INCOMING_REGNUM, while the register -+number as seen by the calling function is STATIC_CHAIN_REGNUM. If -+these registers are the same, STATIC_CHAIN_INCOMING_REGNUM need -+not be defined. -+ -+The static chain register need not be a fixed register. -+ -+If the static chain is passed in memory, these macros should not be -+defined; instead, the next two macros should be defined. -+*/ -+/* Using r0 */ -+#define STATIC_CHAIN_REGNUM INTERNAL_REGNUM(0) -+ -+ -+/** Eliminating Frame Pointer and Arg Pointer **/ -+ -+/* -+A C expression which is nonzero if a function must have and use a frame -+pointer. This expression is evaluated in the reload pass. If its value is -+nonzero the function will have a frame pointer. -+ -+The expression can in principle examine the current function and decide -+according to the facts, but on most machines the constant 0 or the -+constant 1 suffices. Use 0 when the machine allows code to be generated -+with no frame pointer, and doing so saves some time or space. Use 1 -+when there is no possible advantage to avoiding a frame pointer. -+ -+In certain cases, the compiler does not know how to produce valid code -+without a frame pointer. The compiler recognizes those cases and -+automatically gives the function a frame pointer regardless of what -+FRAME_POINTER_REQUIRED says. You don't need to worry about -+them. -+ -+In a function that does not require a frame pointer, the frame pointer -+register can be allocated for ordinary usage, unless you mark it as a -+fixed register. See FIXED_REGISTERS for more information. -+*/ -+/* We need the frame pointer when compiling for profiling */ -+#define FRAME_POINTER_REQUIRED (current_function_profile) -+ -+/* -+A C statement to store in the variable DEPTH_VAR the difference -+between the frame pointer and the stack pointer values immediately after -+the function prologue. The value would be computed from information -+such as the result of get_frame_size () and the tables of -+registers regs_ever_live and call_used_regs. -+ -+If ELIMINABLE_REGS is defined, this macro will be not be used and -+need not be defined. Otherwise, it must be defined even if -+FRAME_POINTER_REQUIRED is defined to always be true; in that -+case, you may set DEPTH_VAR to anything. -+*/ -+#define INITIAL_FRAME_POINTER_OFFSET(DEPTH_VAR) ((DEPTH_VAR) = get_frame_size()) -+ -+/* -+If defined, this macro specifies a table of register pairs used to -+eliminate unneeded registers that point into the stack frame. If it is not -+defined, the only elimination attempted by the compiler is to replace -+references to the frame pointer with references to the stack pointer. -+ -+The definition of this macro is a list of structure initializations, each -+of which specifies an original and replacement register. -+ -+On some machines, the position of the argument pointer is not known until -+the compilation is completed. In such a case, a separate hard register -+must be used for the argument pointer. This register can be eliminated by -+replacing it with either the frame pointer or the argument pointer, -+depending on whether or not the frame pointer has been eliminated. -+ -+In this case, you might specify: -+ #define ELIMINABLE_REGS \ -+ {{ARG_POINTER_REGNUM, STACK_POINTER_REGNUM}, \ -+ {ARG_POINTER_REGNUM, FRAME_POINTER_REGNUM}, \ -+ {FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM}} -+ -+Note that the elimination of the argument pointer with the stack pointer is -+specified first since that is the preferred elimination. -+*/ -+#define ELIMINABLE_REGS \ -+{ \ -+ { FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM }, \ -+ { ARG_POINTER_REGNUM, STACK_POINTER_REGNUM }, \ -+ { ARG_POINTER_REGNUM, FRAME_POINTER_REGNUM } \ -+} -+ -+/* -+A C expression that returns nonzero if the compiler is allowed to try -+to replace register number FROM with register number -+TO. This macro need only be defined if ELIMINABLE_REGS -+is defined, and will usually be the constant 1, since most of the cases -+preventing register elimination are things that the compiler already -+knows about. -+*/ -+#define CAN_ELIMINATE(FROM, TO) 1 -+ -+/* -+This macro is similar to INITIAL_FRAME_POINTER_OFFSET. It -+specifies the initial difference between the specified pair of -+registers. This macro must be defined if ELIMINABLE_REGS is -+defined. -+*/ -+#define INITIAL_ELIMINATION_OFFSET(FROM, TO, OFFSET) \ -+ ((OFFSET) = avr32_initial_elimination_offset(FROM, TO)) -+ -+/** Passing Function Arguments on the Stack **/ -+ -+ -+/* -+A C expression. If nonzero, push insns will be used to pass -+outgoing arguments. -+If the target machine does not have a push instruction, set it to zero. -+That directs GCC to use an alternate strategy: to -+allocate the entire argument block and then store the arguments into -+it. When PUSH_ARGS is nonzero, PUSH_ROUNDING must be defined too. -+*/ -+#define PUSH_ARGS 1 -+ -+ -+/* -+A C expression that is the number of bytes actually pushed onto the -+stack when an instruction attempts to push NPUSHED bytes. -+ -+On some machines, the definition -+ -+ #define PUSH_ROUNDING(BYTES) (BYTES) -+ -+will suffice. But on other machines, instructions that appear -+to push one byte actually push two bytes in an attempt to maintain -+alignment. Then the definition should be -+ -+ #define PUSH_ROUNDING(BYTES) (((BYTES) + 1) & ~1) -+*/ -+/* Push 4 bytes at the time. */ -+#define PUSH_ROUNDING(NPUSHED) (((NPUSHED) + 3) & ~3) -+ -+/* -+A C expression. If nonzero, the maximum amount of space required for -+outgoing arguments will be computed and placed into the variable -+current_function_outgoing_args_size. No space will be pushed -+onto the stack for each call; instead, the function prologue should -+increase the stack frame size by this amount. -+ -+Setting both PUSH_ARGS and ACCUMULATE_OUTGOING_ARGS is not proper. -+*/ -+#define ACCUMULATE_OUTGOING_ARGS 0 -+ -+ -+ -+ -+/* -+A C expression that should indicate the number of bytes of its own -+arguments that a function pops on returning, or 0 if the -+function pops no arguments and the caller must therefore pop them all -+after the function returns. -+ -+FUNDECL is a C variable whose value is a tree node that describes -+the function in question. Normally it is a node of type -+FUNCTION_DECL that describes the declaration of the function. -+From this you can obtain the DECL_ATTRIBUTES of the function. -+ -+FUNTYPE is a C variable whose value is a tree node that -+describes the function in question. Normally it is a node of type -+FUNCTION_TYPE that describes the data type of the function. -+From this it is possible to obtain the data types of the value and -+arguments (if known). -+ -+When a call to a library function is being considered, FUNDECL -+will contain an identifier node for the library function. Thus, if -+you need to distinguish among various library functions, you can do so -+by their names. Note that ``library function'' in this context means -+a function used to perform arithmetic, whose name is known specially -+in the compiler and was not mentioned in the C code being compiled. -+ -+STACK_SIZE is the number of bytes of arguments passed on the -+stack. If a variable number of bytes is passed, it is zero, and -+argument popping will always be the responsibility of the calling function. -+ -+On the VAX, all functions always pop their arguments, so the definition -+of this macro is STACK_SIZE. On the 68000, using the standard -+calling convention, no functions pop their arguments, so the value of -+the macro is always 0 in this case. But an alternative calling -+convention is available in which functions that take a fixed number of -+arguments pop them but other functions (such as printf) pop -+nothing (the caller pops all). When this convention is in use, -+FUNTYPE is examined to determine whether a function takes a fixed -+number of arguments. -+*/ -+#define RETURN_POPS_ARGS(FUNDECL, FUNTYPE, STACK_SIZE) 0 -+ -+ -+/*Return true if this function can we use a single return instruction*/ -+#define USE_RETURN_INSN(ISCOND) avr32_use_return_insn(ISCOND) -+ -+/* -+A C expression that should indicate the number of bytes a call sequence -+pops off the stack. It is added to the value of RETURN_POPS_ARGS -+when compiling a function call. -+ -+CUM is the variable in which all arguments to the called function -+have been accumulated. -+ -+On certain architectures, such as the SH5, a call trampoline is used -+that pops certain registers off the stack, depending on the arguments -+that have been passed to the function. Since this is a property of the -+call site, not of the called function, RETURN_POPS_ARGS is not -+appropriate. -+*/ -+#define CALL_POPS_ARGS(CUM) 0 -+ -+/* Passing Arguments in Registers */ -+ -+/* -+A C expression that controls whether a function argument is passed -+in a register, and which register. -+ -+The arguments are CUM, which summarizes all the previous -+arguments; MODE, the machine mode of the argument; TYPE, -+the data type of the argument as a tree node or 0 if that is not known -+(which happens for C support library functions); and NAMED, -+which is 1 for an ordinary argument and 0 for nameless arguments that -+correspond to '...' in the called function's prototype. -+TYPE can be an incomplete type if a syntax error has previously -+occurred. -+ -+The value of the expression is usually either a reg RTX for the -+hard register in which to pass the argument, or zero to pass the -+argument on the stack. -+ -+For machines like the VAX and 68000, where normally all arguments are -+pushed, zero suffices as a definition. -+ -+The value of the expression can also be a parallel RTX. This is -+used when an argument is passed in multiple locations. The mode of the -+of the parallel should be the mode of the entire argument. The -+parallel holds any number of expr_list pairs; each one -+describes where part of the argument is passed. In each -+expr_list the first operand must be a reg RTX for the hard -+register in which to pass this part of the argument, and the mode of the -+register RTX indicates how large this part of the argument is. The -+second operand of the expr_list is a const_int which gives -+the offset in bytes into the entire argument of where this part starts. -+As a special exception the first expr_list in the parallel -+RTX may have a first operand of zero. This indicates that the entire -+argument is also stored on the stack. -+ -+The last time this macro is called, it is called with MODE == VOIDmode, -+and its result is passed to the call or call_value -+pattern as operands 2 and 3 respectively. -+ -+The usual way to make the ISO library 'stdarg.h' work on a machine -+where some arguments are usually passed in registers, is to cause -+nameless arguments to be passed on the stack instead. This is done -+by making FUNCTION_ARG return 0 whenever NAMED is 0. -+ -+You may use the macro MUST_PASS_IN_STACK (MODE, TYPE) -+in the definition of this macro to determine if this argument is of a -+type that must be passed in the stack. If REG_PARM_STACK_SPACE -+is not defined and FUNCTION_ARG returns nonzero for such an -+argument, the compiler will abort. If REG_PARM_STACK_SPACE is -+defined, the argument will be computed in the stack and then loaded into -+a register. */ -+ -+#define FUNCTION_ARG(CUM, MODE, TYPE, NAMED) \ -+ avr32_function_arg(&(CUM), MODE, TYPE, NAMED) -+ -+ -+ -+ -+/* -+A C type for declaring a variable that is used as the first argument of -+FUNCTION_ARG and other related values. For some target machines, -+the type int suffices and can hold the number of bytes of -+argument so far. -+ -+There is no need to record in CUMULATIVE_ARGS anything about the -+arguments that have been passed on the stack. The compiler has other -+variables to keep track of that. For target machines on which all -+arguments are passed on the stack, there is no need to store anything in -+CUMULATIVE_ARGS; however, the data structure must exist and -+should not be empty, so use int. -+*/ -+typedef struct avr32_args -+{ -+ /* Index representing the argument register the current function argument -+ will occupy */ -+ int index; -+ /* A mask with bits representing the argument registers: if a bit is set -+ then this register is used for an arguemnt */ -+ int used_index; -+ /* TRUE if this function has anonymous arguments */ -+ int uses_anonymous_args; -+ /* The size in bytes of the named arguments pushed on the stack */ -+ int stack_pushed_args_size; -+ /* Set to true if this function needs a Return Value Pointer */ -+ int use_rvp; -+ -+} CUMULATIVE_ARGS; -+ -+ -+#define FIRST_CUM_REG_INDEX 0 -+#define LAST_CUM_REG_INDEX 4 -+#define GET_REG_INDEX(CUM) ((CUM)->index) -+#define SET_REG_INDEX(CUM, INDEX) ((CUM)->index = (INDEX)); -+#define GET_USED_INDEX(CUM, INDEX) ((CUM)->used_index & (1 << (INDEX))) -+#define SET_USED_INDEX(CUM, INDEX) \ -+ do \ -+ { \ -+ if (INDEX >= 0) \ -+ (CUM)->used_index |= (1 << (INDEX)); \ -+ } \ -+ while (0) -+#define SET_INDEXES_UNUSED(CUM) ((CUM)->used_index = 0) -+ -+ -+/* -+ A C statement (sans semicolon) for initializing the variable cum for the -+ state at the beginning of the argument list. The variable has type -+ CUMULATIVE_ARGS. The value of FNTYPE is the tree node for the data type of -+ the function which will receive the args, or 0 if the args are to a compiler -+ support library function. For direct calls that are not libcalls, FNDECL -+ contain the declaration node of the function. FNDECL is also set when -+ INIT_CUMULATIVE_ARGS is used to find arguments for the function being -+ compiled. N_NAMED_ARGS is set to the number of named arguments, including a -+ structure return address if it is passed as a parameter, when making a call. -+ When processing incoming arguments, N_NAMED_ARGS is set to -1. -+ -+ When processing a call to a compiler support library function, LIBNAME -+ identifies which one. It is a symbol_ref rtx which contains the name of the -+ function, as a string. LIBNAME is 0 when an ordinary C function call is -+ being processed. Thus, each time this macro is called, either LIBNAME or -+ FNTYPE is nonzero, but never both of them at once. -+*/ -+#define INIT_CUMULATIVE_ARGS(CUM, FNTYPE, LIBNAME, FNDECL, N_NAMED_ARGS) \ -+ avr32_init_cumulative_args(&(CUM), FNTYPE, LIBNAME, FNDECL) -+ -+ -+/* -+A C statement (sans semicolon) to update the summarizer variable -+CUM to advance past an argument in the argument list. The -+values MODE, TYPE and NAMED describe that argument. -+Once this is done, the variable CUM is suitable for analyzing -+the following argument with FUNCTION_ARG, etc. -+ -+This macro need not do anything if the argument in question was passed -+on the stack. The compiler knows how to track the amount of stack space -+used for arguments without any special help. -+*/ -+#define FUNCTION_ARG_ADVANCE(CUM, MODE, TYPE, NAMED) \ -+ avr32_function_arg_advance(&(CUM), MODE, TYPE, NAMED) -+ -+/* -+If defined, a C expression which determines whether, and in which direction, -+to pad out an argument with extra space. The value should be of type -+enum direction: either 'upward' to pad above the argument, -+'downward' to pad below, or 'none' to inhibit padding. -+ -+The amount of padding is always just enough to reach the next -+multiple of FUNCTION_ARG_BOUNDARY; this macro does not control -+it. -+ -+This macro has a default definition which is right for most systems. -+For little-endian machines, the default is to pad upward. For -+big-endian machines, the default is to pad downward for an argument of -+constant size shorter than an int, and upward otherwise. -+*/ -+#define FUNCTION_ARG_PADDING(MODE, TYPE) \ -+ avr32_function_arg_padding(MODE, TYPE) -+ -+/* -+ Specify padding for the last element of a block move between registers -+ and memory. First is nonzero if this is the only element. Defining -+ this macro allows better control of register function parameters on -+ big-endian machines, without using PARALLEL rtl. In particular, -+ MUST_PASS_IN_STACK need not test padding and mode of types in registers, -+ as there is no longer a "wrong" part of a register; For example, a three -+ byte aggregate may be passed in the high part of a register if so required. -+*/ -+#define BLOCK_REG_PADDING(MODE, TYPE, FIRST) \ -+ avr32_function_arg_padding(MODE, TYPE) -+ -+/* -+If defined, a C expression which determines whether the default -+implementation of va_arg will attempt to pad down before reading the -+next argument, if that argument is smaller than its aligned space as -+controlled by PARM_BOUNDARY. If this macro is not defined, all such -+arguments are padded down if BYTES_BIG_ENDIAN is true. -+*/ -+#define PAD_VARARGS_DOWN \ -+ (FUNCTION_ARG_PADDING (TYPE_MODE (type), type) == downward) -+ -+ -+/* -+A C expression that is nonzero if REGNO is the number of a hard -+register in which function arguments are sometimes passed. This does -+not include implicit arguments such as the static chain and -+the structure-value address. On many machines, no registers can be -+used for this purpose since all function arguments are pushed on the -+stack. -+*/ -+/* -+ Use r8 - r12 for function arguments. -+*/ -+#define FUNCTION_ARG_REGNO_P(REGNO) \ -+ (REGNO >= 3 && REGNO <= 7) -+ -+/* Number of registers used for passing function arguments */ -+#define NUM_ARG_REGS 5 -+ -+/* -+If defined, the order in which arguments are loaded into their -+respective argument registers is reversed so that the last -+argument is loaded first. This macro only affects arguments -+passed in registers. -+*/ -+/* #define LOAD_ARGS_REVERSED */ -+ -+/** How Scalar Function Values Are Returned **/ -+ -+/* AVR32 is using r12 as return register. */ -+#define RET_REGISTER (15 - 12) -+ -+ -+/* -+A C expression to create an RTX representing the place where a library -+function returns a value of mode MODE. If the precise function -+being called is known, FUNC is a tree node -+(FUNCTION_DECL) for it; otherwise, func is a null -+pointer. This makes it possible to use a different value-returning -+convention for specific functions when all their calls are -+known. -+ -+Note that "library function" in this context means a compiler -+support routine, used to perform arithmetic, whose name is known -+specially by the compiler and was not mentioned in the C code being -+compiled. -+ -+The definition of LIBRARY_VALUE need not be concerned aggregate -+data types, because none of the library functions returns such types. -+*/ -+#define LIBCALL_VALUE(MODE) avr32_libcall_value(MODE) -+ -+/* -+A C expression that is nonzero if REGNO is the number of a hard -+register in which the values of called function may come back. -+ -+A register whose use for returning values is limited to serving as the -+second of a pair (for a value of type double, say) need not be -+recognized by this macro. So for most machines, this definition -+suffices: -+ #define FUNCTION_VALUE_REGNO_P(N) ((N) == 0) -+ -+If the machine has register windows, so that the caller and the called -+function use different registers for the return value, this macro -+should recognize only the caller's register numbers. -+*/ -+/* -+ When returning a value of mode DImode, r11:r10 is used, else r12 is used. -+*/ -+#define FUNCTION_VALUE_REGNO_P(REGNO) ((REGNO) == RET_REGISTER \ -+ || (REGNO) == INTERNAL_REGNUM(11)) -+ -+ -+/** How Large Values Are Returned **/ -+ -+ -+/* -+Define this macro to be 1 if all structure and union return values must be -+in memory. Since this results in slower code, this should be defined -+only if needed for compatibility with other compilers or with an ABI. -+If you define this macro to be 0, then the conventions used for structure -+and union return values are decided by the RETURN_IN_MEMORY macro. -+ -+If not defined, this defaults to the value 1. -+*/ -+#define DEFAULT_PCC_STRUCT_RETURN 0 -+ -+ -+ -+ -+/** Generating Code for Profiling **/ -+ -+/* -+A C statement or compound statement to output to FILE some -+assembler code to call the profiling subroutine mcount. -+ -+The details of how mcount expects to be called are determined by -+your operating system environment, not by GCC. To figure them out, -+compile a small program for profiling using the system's installed C -+compiler and look at the assembler code that results. -+ -+Older implementations of mcount expect the address of a counter -+variable to be loaded into some register. The name of this variable is -+'LP' followed by the number LABELNO, so you would generate -+the name using 'LP%d' in a fprintf. -+*/ -+/* ToDo: fixme */ -+#ifndef FUNCTION_PROFILER -+#define FUNCTION_PROFILER(FILE, LABELNO) \ -+ fprintf((FILE), "/* profiler %d */", (LABELNO)) -+#endif -+ -+ -+/***************************************************************************** -+ * Trampolines for Nested Functions * -+ *****************************************************************************/ -+ -+/* -+A C statement to output, on the stream FILE, assembler code for a -+block of data that contains the constant parts of a trampoline. This -+code should not include a label - the label is taken care of -+automatically. -+ -+If you do not define this macro, it means no template is needed -+for the target. Do not define this macro on systems where the block move -+code to copy the trampoline into place would be larger than the code -+to generate it on the spot. -+*/ -+/* ToDo: correct? */ -+#define TRAMPOLINE_TEMPLATE(FILE) avr32_trampoline_template(FILE); -+ -+ -+/* -+A C expression for the size in bytes of the trampoline, as an integer. -+*/ -+/* ToDo: fixme */ -+#define TRAMPOLINE_SIZE 0x0C -+ -+/* -+Alignment required for trampolines, in bits. -+ -+If you don't define this macro, the value of BIGGEST_ALIGNMENT -+is used for aligning trampolines. -+*/ -+#define TRAMPOLINE_ALIGNMENT 16 -+ -+/* -+A C statement to initialize the variable parts of a trampoline. -+ADDR is an RTX for the address of the trampoline; FNADDR is -+an RTX for the address of the nested function; STATIC_CHAIN is an -+RTX for the static chain value that should be passed to the function -+when it is called. -+*/ -+#define INITIALIZE_TRAMPOLINE(ADDR, FNADDR, STATIC_CHAIN) \ -+ avr32_initialize_trampoline(ADDR, FNADDR, STATIC_CHAIN) -+ -+ -+/****************************************************************************** -+ * Implicit Calls to Library Routines -+ *****************************************************************************/ -+ -+/* Tail calling. */ -+ -+/* A C expression that evaluates to true if it is ok to perform a sibling -+ call to DECL. */ -+#define FUNCTION_OK_FOR_SIBCALL(DECL) 0 -+ -+#define OVERRIDE_OPTIONS avr32_override_options () -+ -+#define OPTIMIZATION_OPTIONS(LEVEL, SIZE) avr32_optimization_options (LEVEL, SIZE) -+ -+ -+/****************************************************************************** -+ * Addressing Modes -+ *****************************************************************************/ -+ -+/* -+A C expression that is nonzero if the machine supports pre-increment, -+pre-decrement, post-increment, or post-decrement addressing respectively. -+*/ -+/* -+ AVR32 supports Rp++ and --Rp -+*/ -+#define HAVE_PRE_INCREMENT 0 -+#define HAVE_PRE_DECREMENT 1 -+#define HAVE_POST_INCREMENT 1 -+#define HAVE_POST_DECREMENT 0 -+ -+/* -+A C expression that is nonzero if the machine supports pre- or -+post-address side-effect generation involving constants other than -+the size of the memory operand. -+*/ -+#define HAVE_PRE_MODIFY_DISP 0 -+#define HAVE_POST_MODIFY_DISP 0 -+ -+/* -+A C expression that is nonzero if the machine supports pre- or -+post-address side-effect generation involving a register displacement. -+*/ -+#define HAVE_PRE_MODIFY_REG 0 -+#define HAVE_POST_MODIFY_REG 0 -+ -+/* -+A C expression that is 1 if the RTX X is a constant which -+is a valid address. On most machines, this can be defined as -+CONSTANT_P (X), but a few machines are more restrictive -+in which constant addresses are supported. -+ -+CONSTANT_P accepts integer-values expressions whose values are -+not explicitly known, such as symbol_ref, label_ref, and -+high expressions and const arithmetic expressions, in -+addition to const_int and const_double expressions. -+*/ -+#define CONSTANT_ADDRESS_P(X) CONSTANT_P(X) -+ -+/* -+A number, the maximum number of registers that can appear in a valid -+memory address. Note that it is up to you to specify a value equal to -+the maximum number that GO_IF_LEGITIMATE_ADDRESS would ever -+accept. -+*/ -+#define MAX_REGS_PER_ADDRESS 2 -+ -+/* -+A C compound statement with a conditional goto LABEL; -+executed if X (an RTX) is a legitimate memory address on the -+target machine for a memory operand of mode MODE. -+ -+It usually pays to define several simpler macros to serve as -+subroutines for this one. Otherwise it may be too complicated to -+understand. -+ -+This macro must exist in two variants: a strict variant and a -+non-strict one. The strict variant is used in the reload pass. It -+must be defined so that any pseudo-register that has not been -+allocated a hard register is considered a memory reference. In -+contexts where some kind of register is required, a pseudo-register -+with no hard register must be rejected. -+ -+The non-strict variant is used in other passes. It must be defined to -+accept all pseudo-registers in every context where some kind of -+register is required. -+ -+Compiler source files that want to use the strict variant of this -+macro define the macro REG_OK_STRICT. You should use an -+#ifdef REG_OK_STRICT conditional to define the strict variant -+in that case and the non-strict variant otherwise. -+ -+Subroutines to check for acceptable registers for various purposes (one -+for base registers, one for index registers, and so on) are typically -+among the subroutines used to define GO_IF_LEGITIMATE_ADDRESS. -+Then only these subroutine macros need have two variants; the higher -+levels of macros may be the same whether strict or not. -+ -+Normally, constant addresses which are the sum of a symbol_ref -+and an integer are stored inside a const RTX to mark them as -+constant. Therefore, there is no need to recognize such sums -+specifically as legitimate addresses. Normally you would simply -+recognize any const as legitimate. -+ -+Usually PRINT_OPERAND_ADDRESS is not prepared to handle constant -+sums that are not marked with const. It assumes that a naked -+plus indicates indexing. If so, then you must reject such -+naked constant sums as illegitimate addresses, so that none of them will -+be given to PRINT_OPERAND_ADDRESS. -+ -+On some machines, whether a symbolic address is legitimate depends on -+the section that the address refers to. On these machines, define the -+macro ENCODE_SECTION_INFO to store the information into the -+symbol_ref, and then check for it here. When you see a -+const, you will have to look inside it to find the -+symbol_ref in order to determine the section. -+ -+The best way to modify the name string is by adding text to the -+beginning, with suitable punctuation to prevent any ambiguity. Allocate -+the new name in saveable_obstack. You will have to modify -+ASM_OUTPUT_LABELREF to remove and decode the added text and -+output the name accordingly, and define STRIP_NAME_ENCODING to -+access the original name string. -+ -+You can check the information stored here into the symbol_ref in -+the definitions of the macros GO_IF_LEGITIMATE_ADDRESS and -+PRINT_OPERAND_ADDRESS. -+*/ -+#ifdef REG_OK_STRICT -+# define GO_IF_LEGITIMATE_ADDRESS(MODE, X, LABEL) \ -+ do \ -+ { \ -+ if (avr32_legitimate_address(MODE, X, 1)) \ -+ goto LABEL; \ -+ } \ -+ while (0) -+#else -+# define GO_IF_LEGITIMATE_ADDRESS(MODE, X, LABEL) \ -+ do \ -+ { \ -+ if (avr32_legitimate_address(MODE, X, 0)) \ -+ goto LABEL; \ -+ } \ -+ while (0) -+#endif -+ -+ -+ -+/* -+A C compound statement that attempts to replace X with a valid -+memory address for an operand of mode MODE. win will be a -+C statement label elsewhere in the code; the macro definition may use -+ -+ GO_IF_LEGITIMATE_ADDRESS (MODE, X, WIN); -+ -+to avoid further processing if the address has become legitimate. -+ -+X will always be the result of a call to break_out_memory_refs, -+and OLDX will be the operand that was given to that function to produce -+X. -+ -+The code generated by this macro should not alter the substructure of -+X. If it transforms X into a more legitimate form, it -+should assign X (which will always be a C variable) a new value. -+ -+It is not necessary for this macro to come up with a legitimate -+address. The compiler has standard ways of doing so in all cases. In -+fact, it is safe for this macro to do nothing. But often a -+machine-dependent strategy can generate better code. -+*/ -+#define LEGITIMIZE_ADDRESS(X, OLDX, MODE, WIN) \ -+ do \ -+ { \ -+ if (GET_CODE(X) == PLUS \ -+ && GET_CODE(XEXP(X, 0)) == REG \ -+ && GET_CODE(XEXP(X, 1)) == CONST_INT \ -+ && !CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(X, 1)), \ -+ 'K', "Ks16")) \ -+ { \ -+ rtx index = force_reg(SImode, XEXP(X, 1)); \ -+ X = gen_rtx_PLUS( SImode, XEXP(X, 0), index); \ -+ } \ -+ GO_IF_LEGITIMATE_ADDRESS(MODE, X, WIN); \ -+ } \ -+ while(0) -+ -+ -+/* -+A C statement or compound statement with a conditional -+goto LABEL; executed if memory address X (an RTX) can have -+different meanings depending on the machine mode of the memory -+reference it is used for or if the address is valid for some modes -+but not others. -+ -+Autoincrement and autodecrement addresses typically have mode-dependent -+effects because the amount of the increment or decrement is the size -+of the operand being addressed. Some machines have other mode-dependent -+addresses. Many RISC machines have no mode-dependent addresses. -+ -+You may assume that ADDR is a valid address for the machine. -+*/ -+#define GO_IF_MODE_DEPENDENT_ADDRESS(ADDR, LABEL) \ -+ do \ -+ { \ -+ if (GET_CODE (ADDR) == POST_INC \ -+ || GET_CODE (ADDR) == PRE_DEC) \ -+ goto LABEL; \ -+ } \ -+ while (0) -+ -+/* -+A C expression that is nonzero if X is a legitimate constant for -+an immediate operand on the target machine. You can assume that -+X satisfies CONSTANT_P, so you need not check this. In fact, -+'1' is a suitable definition for this macro on machines where -+anything CONSTANT_P is valid. -+*/ -+#define LEGITIMATE_CONSTANT_P(X) avr32_legitimate_constant_p(X) -+ -+ -+/****************************************************************************** -+ * Condition Code Status -+ *****************************************************************************/ -+ -+#define HAVE_conditional_move 1 -+ -+/* -+C code for a data type which is used for declaring the mdep -+component of cc_status. It defaults to int. -+ -+This macro is not used on machines that do not use cc0. -+*/ -+ -+typedef struct -+{ -+ int flags; -+ rtx value; -+ int fpflags; -+ rtx fpvalue; -+} avr32_status_reg; -+ -+ -+#define CC_STATUS_MDEP avr32_status_reg -+ -+/* -+A C expression to initialize the mdep field to "empty". -+The default definition does nothing, since most machines don't use -+the field anyway. If you want to use the field, you should probably -+define this macro to initialize it. -+ -+This macro is not used on machines that do not use cc0. -+*/ -+ -+#define CC_STATUS_MDEP_INIT \ -+ (cc_status.mdep.flags = CC_NONE , cc_status.mdep.value = 0) -+ -+#define FPCC_STATUS_INIT \ -+ (cc_status.mdep.fpflags = CC_NONE , cc_status.mdep.fpvalue = 0) -+ -+/* -+A C compound statement to set the components of cc_status -+appropriately for an insn INSN whose body is EXP. It is -+this macro's responsibility to recognize insns that set the condition -+code as a byproduct of other activity as well as those that explicitly -+set (cc0). -+ -+This macro is not used on machines that do not use cc0. -+ -+If there are insns that do not set the condition code but do alter -+other machine registers, this macro must check to see whether they -+invalidate the expressions that the condition code is recorded as -+reflecting. For example, on the 68000, insns that store in address -+registers do not set the condition code, which means that usually -+NOTICE_UPDATE_CC can leave cc_status unaltered for such -+insns. But suppose that the previous insn set the condition code -+based on location 'a4@@(102)' and the current insn stores a new -+value in 'a4'. Although the condition code is not changed by -+this, it will no longer be true that it reflects the contents of -+'a4@@(102)'. Therefore, NOTICE_UPDATE_CC must alter -+cc_status in this case to say that nothing is known about the -+condition code value. -+ -+The definition of NOTICE_UPDATE_CC must be prepared to deal -+with the results of peephole optimization: insns whose patterns are -+parallel RTXs containing various reg, mem or -+constants which are just the operands. The RTL structure of these -+insns is not sufficient to indicate what the insns actually do. What -+NOTICE_UPDATE_CC should do when it sees one is just to run -+CC_STATUS_INIT. -+ -+A possible definition of NOTICE_UPDATE_CC is to call a function -+that looks at an attribute (see Insn Attributes) named, for example, -+'cc'. This avoids having detailed information about patterns in -+two places, the 'md' file and in NOTICE_UPDATE_CC. -+*/ -+ -+#define NOTICE_UPDATE_CC(EXP, INSN) avr32_notice_update_cc(EXP, INSN) -+ -+ -+ -+ -+/****************************************************************************** -+ * Describing Relative Costs of Operations -+ *****************************************************************************/ -+ -+ -+ -+/* -+A C expression for the cost of moving data of mode MODE from a -+register in class FROM to one in class TO. The classes are -+expressed using the enumeration values such as GENERAL_REGS. A -+value of 2 is the default; other values are interpreted relative to -+that. -+ -+It is not required that the cost always equal 2 when FROM is the -+same as TO; on some machines it is expensive to move between -+registers if they are not general registers. -+ -+If reload sees an insn consisting of a single set between two -+hard registers, and if REGISTER_MOVE_COST applied to their -+classes returns a value of 2, reload does not check to ensure that the -+constraints of the insn are met. Setting a cost of other than 2 will -+allow reload to verify that the constraints are met. You should do this -+if the movm pattern's constraints do not allow such copying. -+*/ -+#define REGISTER_MOVE_COST(MODE, FROM, TO) \ -+ ((GET_MODE_SIZE(MODE) <= 4) ? 2: \ -+ (GET_MODE_SIZE(MODE) <= 8) ? 3: \ -+ 4) -+ -+/* -+A C expression for the cost of moving data of mode MODE between a -+register of class CLASS and memory; IN is zero if the value -+is to be written to memory, nonzero if it is to be read in. This cost -+is relative to those in REGISTER_MOVE_COST. If moving between -+registers and memory is more expensive than between two registers, you -+should define this macro to express the relative cost. -+ -+If you do not define this macro, GCC uses a default cost of 4 plus -+the cost of copying via a secondary reload register, if one is -+needed. If your machine requires a secondary reload register to copy -+between memory and a register of CLASS but the reload mechanism is -+more complex than copying via an intermediate, define this macro to -+reflect the actual cost of the move. -+ -+GCC defines the function memory_move_secondary_cost if -+secondary reloads are needed. It computes the costs due to copying via -+a secondary register. If your machine copies from memory using a -+secondary register in the conventional way but the default base value of -+4 is not correct for your machine, define this macro to add some other -+value to the result of that function. The arguments to that function -+are the same as to this macro. -+*/ -+/* -+ Memory moves are costly -+*/ -+#define MEMORY_MOVE_COST(MODE, CLASS, IN) \ -+ (((IN) ? ((GET_MODE_SIZE(MODE) < 4) ? 4 : \ -+ (GET_MODE_SIZE(MODE) > 8) ? 6 : \ -+ 3) \ -+ : ((GET_MODE_SIZE(MODE) > 8) ? 6 : 3))) -+ -+/* -+A C expression for the cost of a branch instruction. A value of 1 is -+the default; other values are interpreted relative to that. -+*/ -+ /* Try to use conditionals as much as possible */ -+#define BRANCH_COST (TARGET_BRANCH_PRED ? 3 : 4) -+ -+/*A C expression for the maximum number of instructions to execute via conditional -+ execution instructions instead of a branch. A value of BRANCH_COST+1 is the default -+ if the machine does not use cc0, and 1 if it does use cc0.*/ -+#define MAX_CONDITIONAL_EXECUTE 4 -+ -+/* -+Define this macro as a C expression which is nonzero if accessing less -+than a word of memory (i.e.: a char or a short) is no -+faster than accessing a word of memory, i.e., if such access -+require more than one instruction or if there is no difference in cost -+between byte and (aligned) word loads. -+ -+When this macro is not defined, the compiler will access a field by -+finding the smallest containing object; when it is defined, a fullword -+load will be used if alignment permits. Unless bytes accesses are -+faster than word accesses, using word accesses is preferable since it -+may eliminate subsequent memory access if subsequent accesses occur to -+other fields in the same word of the structure, but to different bytes. -+*/ -+#define SLOW_BYTE_ACCESS 1 -+ -+ -+/* -+Define this macro if it is as good or better to call a constant -+function address than to call an address kept in a register. -+*/ -+#define NO_FUNCTION_CSE -+ -+ -+/****************************************************************************** -+ * Adjusting the Instruction Scheduler -+ *****************************************************************************/ -+ -+/***************************************************************************** -+ * Dividing the Output into Sections (Texts, Data, ...) * -+ *****************************************************************************/ -+ -+/* -+A C expression whose value is a string, including spacing, containing the -+assembler operation that should precede instructions and read-only data. -+Normally "\t.text" is right. -+*/ -+#define TEXT_SECTION_ASM_OP "\t.text" -+/* -+A C statement that switches to the default section containing instructions. -+Normally this is not needed, as simply defining TEXT_SECTION_ASM_OP -+is enough. The MIPS port uses this to sort all functions after all data -+declarations. -+*/ -+/* #define TEXT_SECTION */ -+ -+/* -+A C expression whose value is a string, including spacing, containing the -+assembler operation to identify the following data as writable initialized -+data. Normally "\t.data" is right. -+*/ -+#define DATA_SECTION_ASM_OP "\t.data" -+ -+/* -+If defined, a C expression whose value is a string, including spacing, -+containing the assembler operation to identify the following data as -+shared data. If not defined, DATA_SECTION_ASM_OP will be used. -+*/ -+ -+/* -+A C expression whose value is a string, including spacing, containing -+the assembler operation to identify the following data as read-only -+initialized data. -+*/ -+#undef READONLY_DATA_SECTION_ASM_OP -+#define READONLY_DATA_SECTION_ASM_OP \ -+ ((TARGET_USE_RODATA_SECTION) ? \ -+ "\t.section\t.rodata" : \ -+ TEXT_SECTION_ASM_OP ) -+ -+ -+/* -+If defined, a C expression whose value is a string, including spacing, -+containing the assembler operation to identify the following data as -+uninitialized global data. If not defined, and neither -+ASM_OUTPUT_BSS nor ASM_OUTPUT_ALIGNED_BSS are defined, -+uninitialized global data will be output in the data section if -+-fno-common is passed, otherwise ASM_OUTPUT_COMMON will be -+used. -+*/ -+#define BSS_SECTION_ASM_OP "\t.section\t.bss" -+ -+/* -+If defined, a C expression whose value is a string, including spacing, -+containing the assembler operation to identify the following data as -+uninitialized global shared data. If not defined, and -+BSS_SECTION_ASM_OP is, the latter will be used. -+*/ -+/*#define SHARED_BSS_SECTION_ASM_OP "\trseg\tshared_bbs_section:data:noroot(0)\n"*/ -+/* -+If defined, a C expression whose value is a string, including spacing, -+containing the assembler operation to identify the following data as -+initialization code. If not defined, GCC will assume such a section does -+not exist. -+*/ -+#undef INIT_SECTION_ASM_OP -+#define INIT_SECTION_ASM_OP "\t.section\t.init" -+ -+/* -+If defined, a C expression whose value is a string, including spacing, -+containing the assembler operation to identify the following data as -+finalization code. If not defined, GCC will assume such a section does -+not exist. -+*/ -+#undef FINI_SECTION_ASM_OP -+#define FINI_SECTION_ASM_OP "\t.section\t.fini" -+ -+/* -+If defined, an ASM statement that switches to a different section -+via SECTION_OP, calls FUNCTION, and switches back to -+the text section. This is used in crtstuff.c if -+INIT_SECTION_ASM_OP or FINI_SECTION_ASM_OP to calls -+to initialization and finalization functions from the init and fini -+sections. By default, this macro uses a simple function call. Some -+ports need hand-crafted assembly code to avoid dependencies on -+registers initialized in the function prologue or to ensure that -+constant pools don't end up too far way in the text section. -+*/ -+#define CRT_CALL_STATIC_FUNCTION(SECTION_OP, FUNC) \ -+ asm ( SECTION_OP "\n" \ -+ "mcall r6[" USER_LABEL_PREFIX #FUNC "@got]\n" \ -+ TEXT_SECTION_ASM_OP); -+ -+ -+/* -+Define this macro to be an expression with a nonzero value if jump -+tables (for tablejump insns) should be output in the text -+section, along with the assembler instructions. Otherwise, the -+readonly data section is used. -+ -+This macro is irrelevant if there is no separate readonly data section. -+*/ -+/* Put jump tables in text section if we have caches. Otherwise assume that -+ loading data from code memory is slow. */ -+#define JUMP_TABLES_IN_TEXT_SECTION \ -+ (TARGET_CACHES ? 1 : 0) -+ -+ -+/****************************************************************************** -+ * Position Independent Code (PIC) -+ *****************************************************************************/ -+ -+#ifndef AVR32_ALWAYS_PIC -+#define AVR32_ALWAYS_PIC 0 -+#endif -+ -+/* GOT is set to r6 */ -+#define PIC_OFFSET_TABLE_REGNUM INTERNAL_REGNUM(6) -+ -+/* -+A C expression that is nonzero if X is a legitimate immediate -+operand on the target machine when generating position independent code. -+You can assume that X satisfies CONSTANT_P, so you need not -+check this. You can also assume flag_pic is true, so you need not -+check it either. You need not define this macro if all constants -+(including SYMBOL_REF) can be immediate operands when generating -+position independent code. -+*/ -+/* We can't directly access anything that contains a symbol, -+ nor can we indirect via the constant pool. */ -+#define LEGITIMATE_PIC_OPERAND_P(X) avr32_legitimate_pic_operand_p(X) -+ -+ -+/* We need to know when we are making a constant pool; this determines -+ whether data needs to be in the GOT or can be referenced via a GOT -+ offset. */ -+extern int making_const_table; -+ -+/****************************************************************************** -+ * Defining the Output Assembler Language -+ *****************************************************************************/ -+ -+ -+/* -+A C string constant describing how to begin a comment in the target -+assembler language. The compiler assumes that the comment will end at -+the end of the line. -+*/ -+#define ASM_COMMENT_START "# " -+ -+/* -+A C string constant for text to be output before each asm -+statement or group of consecutive ones. Normally this is -+"#APP", which is a comment that has no effect on most -+assemblers but tells the GNU assembler that it must check the lines -+that follow for all valid assembler constructs. -+*/ -+#undef ASM_APP_ON -+#define ASM_APP_ON "#APP\n" -+ -+/* -+A C string constant for text to be output after each asm -+statement or group of consecutive ones. Normally this is -+"#NO_APP", which tells the GNU assembler to resume making the -+time-saving assumptions that are valid for ordinary compiler output. -+*/ -+#undef ASM_APP_OFF -+#define ASM_APP_OFF "#NO_APP\n" -+ -+ -+ -+#define FILE_ASM_OP "\t.file\n" -+#define IDENT_ASM_OP "\t.ident\t" -+#define SET_ASM_OP "\t.set\t" -+ -+ -+/* -+ * Output assembly directives to switch to section name. The section -+ * should have attributes as specified by flags, which is a bit mask -+ * of the SECTION_* flags defined in 'output.h'. If align is nonzero, -+ * it contains an alignment in bytes to be used for the section, -+ * otherwise some target default should be used. Only targets that -+ * must specify an alignment within the section directive need pay -+ * attention to align -- we will still use ASM_OUTPUT_ALIGN. -+ * -+ * NOTE: This one must not be moved to avr32.c -+ */ -+#undef TARGET_ASM_NAMED_SECTION -+#define TARGET_ASM_NAMED_SECTION default_elf_asm_named_section -+ -+ -+/* -+You may define this macro as a C expression. You should define the -+expression to have a nonzero value if GCC should output the constant -+pool for a function before the code for the function, or a zero value if -+GCC should output the constant pool after the function. If you do -+not define this macro, the usual case, GCC will output the constant -+pool before the function. -+*/ -+#define CONSTANT_POOL_BEFORE_FUNCTION 0 -+ -+ -+/* -+Define this macro as a C expression which is nonzero if the constant -+EXP, of type tree, should be output after the code for a -+function. The compiler will normally output all constants before the -+function; you need not define this macro if this is OK. -+*/ -+#define CONSTANT_AFTER_FUNCTION_P(EXP) 1 -+ -+ -+/* -+Define this macro as a C expression which is nonzero if C is -+used as a logical line separator by the assembler. -+ -+If you do not define this macro, the default is that only -+the character ';' is treated as a logical line separator. -+*/ -+#define IS_ASM_LOGICAL_LINE_SEPARATOR(C) ((C) == '\n') -+ -+ -+/** Output of Uninitialized Variables **/ -+ -+/* -+A C statement (sans semicolon) to output to the stdio stream -+STREAM the assembler definition of a common-label named -+NAME whose size is SIZE bytes. The variable ROUNDED -+is the size rounded up to whatever alignment the caller wants. -+ -+Use the expression assemble_name(STREAM, NAME) to -+output the name itself; before and after that, output the additional -+assembler syntax for defining the name, and a newline. -+ -+This macro controls how the assembler definitions of uninitialized -+common global variables are output. -+*/ -+/* -+#define ASM_OUTPUT_COMMON(STREAM, NAME, SIZE, ROUNDED) \ -+ avr32_asm_output_common(STREAM, NAME, SIZE, ROUNDED) -+*/ -+ -+#define ASM_OUTPUT_COMMON(FILE, NAME, SIZE, ROUNDED) \ -+ do \ -+ { \ -+ fputs ("\t.comm ", (FILE)); \ -+ assemble_name ((FILE), (NAME)); \ -+ fprintf ((FILE), ",%d\n", (SIZE)); \ -+ } \ -+ while (0) -+ -+/* -+ * Like ASM_OUTPUT_BSS except takes the required alignment as a -+ * separate, explicit argument. If you define this macro, it is used -+ * in place of ASM_OUTPUT_BSS, and gives you more flexibility in -+ * handling the required alignment of the variable. The alignment is -+ * specified as the number of bits. -+ * -+ * Try to use function asm_output_aligned_bss defined in file varasm.c -+ * when defining this macro. -+ */ -+#define ASM_OUTPUT_ALIGNED_BSS(STREAM, DECL, NAME, SIZE, ALIGNMENT) \ -+ asm_output_aligned_bss (STREAM, DECL, NAME, SIZE, ALIGNMENT) -+ -+/* -+A C statement (sans semicolon) to output to the stdio stream -+STREAM the assembler definition of a local-common-label named -+NAME whose size is SIZE bytes. The variable ROUNDED -+is the size rounded up to whatever alignment the caller wants. -+ -+Use the expression assemble_name(STREAM, NAME) to -+output the name itself; before and after that, output the additional -+assembler syntax for defining the name, and a newline. -+ -+This macro controls how the assembler definitions of uninitialized -+static variables are output. -+*/ -+#define ASM_OUTPUT_LOCAL(FILE, NAME, SIZE, ROUNDED) \ -+ do \ -+ { \ -+ fputs ("\t.lcomm ", (FILE)); \ -+ assemble_name ((FILE), (NAME)); \ -+ fprintf ((FILE), ",%d, %d\n", (SIZE), 2); \ -+ } \ -+ while (0) -+ -+ -+/* -+A C statement (sans semicolon) to output to the stdio stream -+STREAM the assembler definition of a label named NAME. -+Use the expression assemble_name(STREAM, NAME) to -+output the name itself; before and after that, output the additional -+assembler syntax for defining the name, and a newline. -+*/ -+#define ASM_OUTPUT_LABEL(STREAM, NAME) avr32_asm_output_label(STREAM, NAME) -+ -+/* A C string containing the appropriate assembler directive to -+ * specify the size of a symbol, without any arguments. On systems -+ * that use ELF, the default (in 'config/elfos.h') is '"\t.size\t"'; -+ * on other systems, the default is not to define this macro. -+ * -+ * Define this macro only if it is correct to use the default -+ * definitions of ASM_ OUTPUT_SIZE_DIRECTIVE and -+ * ASM_OUTPUT_MEASURED_SIZE for your system. If you need your own -+ * custom definitions of those macros, or if you do not need explicit -+ * symbol sizes at all, do not define this macro. -+ */ -+#define SIZE_ASM_OP "\t.size\t" -+ -+ -+/* -+A C statement (sans semicolon) to output to the stdio stream -+STREAM some commands that will make the label NAME global; -+that is, available for reference from other files. Use the expression -+assemble_name(STREAM, NAME) to output the name -+itself; before and after that, output the additional assembler syntax -+for making that name global, and a newline. -+*/ -+#define GLOBAL_ASM_OP "\t.globl\t" -+ -+ -+ -+/* -+A C expression which evaluates to true if the target supports weak symbols. -+ -+If you don't define this macro, defaults.h provides a default -+definition. If either ASM_WEAKEN_LABEL or ASM_WEAKEN_DECL -+is defined, the default definition is '1'; otherwise, it is -+'0'. Define this macro if you want to control weak symbol support -+with a compiler flag such as -melf. -+*/ -+#define SUPPORTS_WEAK 1 -+ -+/* -+A C statement (sans semicolon) to output to the stdio stream -+STREAM a reference in assembler syntax to a label named -+NAME. This should add '_' to the front of the name, if that -+is customary on your operating system, as it is in most Berkeley Unix -+systems. This macro is used in assemble_name. -+*/ -+#define ASM_OUTPUT_LABELREF(STREAM, NAME) \ -+ avr32_asm_output_labelref(STREAM, NAME) -+ -+ -+ -+/* -+A C expression to assign to OUTVAR (which is a variable of type -+char *) a newly allocated string made from the string -+NAME and the number NUMBER, with some suitable punctuation -+added. Use alloca to get space for the string. -+ -+The string will be used as an argument to ASM_OUTPUT_LABELREF to -+produce an assembler label for an internal static variable whose name is -+NAME. Therefore, the string must be such as to result in valid -+assembler code. The argument NUMBER is different each time this -+macro is executed; it prevents conflicts between similarly-named -+internal static variables in different scopes. -+ -+Ideally this string should not be a valid C identifier, to prevent any -+conflict with the user's own symbols. Most assemblers allow periods -+or percent signs in assembler symbols; putting at least one of these -+between the name and the number will suffice. -+*/ -+#define ASM_FORMAT_PRIVATE_NAME(OUTVAR, NAME, NUMBER) \ -+ do \ -+ { \ -+ (OUTVAR) = (char *) alloca (strlen ((NAME)) + 10); \ -+ sprintf ((OUTVAR), "%s.%d", (NAME), (NUMBER)); \ -+ } \ -+ while (0) -+ -+ -+/** Macros Controlling Initialization Routines **/ -+ -+ -+/* -+If defined, main will not call __main as described above. -+This macro should be defined for systems that control start-up code -+on a symbol-by-symbol basis, such as OSF/1, and should not -+be defined explicitly for systems that support INIT_SECTION_ASM_OP. -+*/ -+/* -+ __main is not defined when debugging. -+*/ -+#define HAS_INIT_SECTION -+ -+ -+/** Output of Assembler Instructions **/ -+ -+/* -+A C initializer containing the assembler's names for the machine -+registers, each one as a C string constant. This is what translates -+register numbers in the compiler into assembler language. -+*/ -+ -+#define REGISTER_NAMES \ -+{ \ -+ "pc", "lr", \ -+ "sp", "r12", \ -+ "r11", "r10", \ -+ "r9", "r8", \ -+ "r7", "r6", \ -+ "r5", "r4", \ -+ "r3", "r2", \ -+ "r1", "r0", \ -+ "f15","f14", \ -+ "f13","f12", \ -+ "f11","f10", \ -+ "f9", "f8", \ -+ "f7", "f6", \ -+ "f5", "f4", \ -+ "f3", "f2", \ -+ "f1", "f0" \ -+} -+ -+/* -+A C compound statement to output to stdio stream STREAM the -+assembler syntax for an instruction operand X. X is an -+RTL expression. -+ -+CODE is a value that can be used to specify one of several ways -+of printing the operand. It is used when identical operands must be -+printed differently depending on the context. CODE comes from -+the '%' specification that was used to request printing of the -+operand. If the specification was just '%digit' then -+CODE is 0; if the specification was '%ltr digit' -+then CODE is the ASCII code for ltr. -+ -+If X is a register, this macro should print the register's name. -+The names can be found in an array reg_names whose type is -+char *[]. reg_names is initialized from REGISTER_NAMES. -+ -+When the machine description has a specification '%punct' -+(a '%' followed by a punctuation character), this macro is called -+with a null pointer for X and the punctuation character for -+CODE. -+*/ -+#define PRINT_OPERAND(STREAM, X, CODE) avr32_print_operand(STREAM, X, CODE) -+ -+/* A C statement to be executed just prior to the output of -+ assembler code for INSN, to modify the extracted operands so -+ they will be output differently. -+ -+ Here the argument OPVEC is the vector containing the operands -+ extracted from INSN, and NOPERANDS is the number of elements of -+ the vector which contain meaningful data for this insn. -+ The contents of this vector are what will be used to convert the insn -+ template into assembler code, so you can change the assembler output -+ by changing the contents of the vector. */ -+#define FINAL_PRESCAN_INSN(INSN, OPVEC, NOPERANDS) \ -+ avr32_final_prescan_insn ((INSN), (OPVEC), (NOPERANDS)) -+ -+/* -+A C expression which evaluates to true if CODE is a valid -+punctuation character for use in the PRINT_OPERAND macro. If -+PRINT_OPERAND_PUNCT_VALID_P is not defined, it means that no -+punctuation characters (except for the standard one, '%') are used -+in this way. -+*/ -+/* -+ 'm' refers to the most significant word in a two-register mode. -+*/ -+#define PRINT_OPERAND_PUNCT_VALID_P(CODE) ((CODE) == 'm' || (CODE) == 'e') -+ -+/* -+A C compound statement to output to stdio stream STREAM the -+assembler syntax for an instruction operand that is a memory reference -+whose address is X. X is an RTL expression. -+ -+On some machines, the syntax for a symbolic address depends on the -+section that the address refers to. On these machines, define the macro -+ENCODE_SECTION_INFO to store the information into the -+symbol_ref, and then check for it here. (see Assembler Format.) -+*/ -+#define PRINT_OPERAND_ADDRESS(STREAM, X) avr32_print_operand_address(STREAM, X) -+ -+ -+/** Output of Dispatch Tables **/ -+ -+/* -+ * A C statement to output to the stdio stream stream an assembler -+ * pseudo-instruction to generate a difference between two -+ * labels. value and rel are the numbers of two internal labels. The -+ * definitions of these labels are output using -+ * (*targetm.asm_out.internal_label), and they must be printed in the -+ * same way here. For example, -+ * -+ * fprintf (stream, "\t.word L%d-L%d\n", -+ * value, rel) -+ * -+ * You must provide this macro on machines where the addresses in a -+ * dispatch table are relative to the table's own address. If defined, -+ * GCC will also use this macro on all machines when producing -+ * PIC. body is the body of the ADDR_DIFF_VEC; it is provided so that -+ * the mode and flags can be read. -+ */ -+#define ASM_OUTPUT_ADDR_DIFF_ELT(STREAM, BODY, VALUE, REL) \ -+ fprintf(STREAM, "\tbral\t%sL%d\n", LOCAL_LABEL_PREFIX, VALUE) -+ -+/* -+This macro should be provided on machines where the addresses -+in a dispatch table are absolute. -+ -+The definition should be a C statement to output to the stdio stream -+STREAM an assembler pseudo-instruction to generate a reference to -+a label. VALUE is the number of an internal label whose -+definition is output using ASM_OUTPUT_INTERNAL_LABEL. -+For example, -+ -+fprintf(STREAM, "\t.word L%d\n", VALUE) -+*/ -+ -+#define ASM_OUTPUT_ADDR_VEC_ELT(STREAM, VALUE) \ -+ fprintf(STREAM, "\t.long %sL%d\n", LOCAL_LABEL_PREFIX, VALUE) -+ -+/** Assembler Commands for Exception Regions */ -+ -+/* ToDo: All of this subsection */ -+ -+/** Assembler Commands for Alignment */ -+ -+ -+/* -+A C statement to output to the stdio stream STREAM an assembler -+command to advance the location counter to a multiple of 2 to the -+POWER bytes. POWER will be a C expression of type int. -+*/ -+#define ASM_OUTPUT_ALIGN(STREAM, POWER) \ -+ do \ -+ { \ -+ if ((POWER) != 0) \ -+ fprintf(STREAM, "\t.align\t%d\n", POWER); \ -+ } \ -+ while (0) -+ -+/* -+Like ASM_OUTPUT_ALIGN, except that the \nop" instruction is used for padding, if -+necessary. -+*/ -+#define ASM_OUTPUT_ALIGN_WITH_NOP(STREAM, POWER) \ -+ fprintf(STREAM, "\t.balignw\t%d, 0xd703\n", (1 << POWER)) -+ -+ -+ -+/****************************************************************************** -+ * Controlling Debugging Information Format -+ *****************************************************************************/ -+ -+/* How to renumber registers for dbx and gdb. */ -+#define DBX_REGISTER_NUMBER(REGNO) ASM_REGNUM (REGNO) -+ -+/* The DWARF 2 CFA column which tracks the return address. */ -+#define DWARF_FRAME_RETURN_COLUMN DWARF_FRAME_REGNUM(LR_REGNUM) -+ -+/* -+Define this macro if GCC should produce dwarf version 2 format -+debugging output in response to the -g option. -+ -+To support optional call frame debugging information, you must also -+define INCOMING_RETURN_ADDR_RTX and either set -+RTX_FRAME_RELATED_P on the prologue insns if you use RTL for the -+prologue, or call dwarf2out_def_cfa and dwarf2out_reg_save -+as appropriate from TARGET_ASM_FUNCTION_PROLOGUE if you don't. -+*/ -+#define DWARF2_DEBUGGING_INFO 1 -+ -+ -+#define DWARF2_ASM_LINE_DEBUG_INFO 1 -+#define DWARF2_FRAME_INFO 1 -+ -+ -+/****************************************************************************** -+ * Miscellaneous Parameters -+ *****************************************************************************/ -+ -+/* ToDo: a lot */ -+ -+/* -+An alias for a machine mode name. This is the machine mode that -+elements of a jump-table should have. -+*/ -+#define CASE_VECTOR_MODE SImode -+ -+/* -+Define this macro to be a C expression to indicate when jump-tables -+should contain relative addresses. If jump-tables never contain -+relative addresses, then you need not define this macro. -+*/ -+#define CASE_VECTOR_PC_RELATIVE 0 -+ -+/* Increase the threshold for using table jumps on the UC arch. */ -+#define CASE_VALUES_THRESHOLD ((avr32_arch->arch_type == ARCH_TYPE_AVR32_UC) ? 7 : 4) -+ -+/* -+The maximum number of bytes that a single instruction can move quickly -+between memory and registers or between two memory locations. -+*/ -+#define MOVE_MAX (2*UNITS_PER_WORD) -+ -+ -+/* A C expression that is nonzero if on this machine the number of bits actually used -+ for the count of a shift operation is equal to the number of bits needed to represent -+ the size of the object being shifted. When this macro is nonzero, the compiler will -+ assume that it is safe to omit a sign-extend, zero-extend, and certain bitwise 'and' -+ instructions that truncates the count of a shift operation. On machines that have -+ instructions that act on bit-fields at variable positions, which may include 'bit test' -+ 378 GNU Compiler Collection (GCC) Internals -+ instructions, a nonzero SHIFT_COUNT_TRUNCATED also enables deletion of truncations -+ of the values that serve as arguments to bit-field instructions. -+ If both types of instructions truncate the count (for shifts) and position (for bit-field -+ operations), or if no variable-position bit-field instructions exist, you should define -+ this macro. -+ However, on some machines, such as the 80386 and the 680x0, truncation only applies -+ to shift operations and not the (real or pretended) bit-field operations. Define SHIFT_ -+ COUNT_TRUNCATED to be zero on such machines. Instead, add patterns to the 'md' file -+ that include the implied truncation of the shift instructions. -+ You need not de ne this macro if it would always have the value of zero. */ -+#define SHIFT_COUNT_TRUNCATED 1 -+ -+/* -+A C expression which is nonzero if on this machine it is safe to -+convert an integer of INPREC bits to one of OUTPREC -+bits (where OUTPREC is smaller than INPREC) by merely -+operating on it as if it had only OUTPREC bits. -+ -+On many machines, this expression can be 1. -+ -+When TRULY_NOOP_TRUNCATION returns 1 for a pair of sizes for -+modes for which MODES_TIEABLE_P is 0, suboptimal code can result. -+If this is the case, making TRULY_NOOP_TRUNCATION return 0 in -+such cases may improve things. -+*/ -+#define TRULY_NOOP_TRUNCATION(OUTPREC, INPREC) 1 -+ -+/* -+An alias for the machine mode for pointers. On most machines, define -+this to be the integer mode corresponding to the width of a hardware -+pointer; SImode on 32-bit machine or DImode on 64-bit machines. -+On some machines you must define this to be one of the partial integer -+modes, such as PSImode. -+ -+The width of Pmode must be at least as large as the value of -+POINTER_SIZE. If it is not equal, you must define the macro -+POINTERS_EXTEND_UNSIGNED to specify how pointers are extended -+to Pmode. -+*/ -+#define Pmode SImode -+ -+/* -+An alias for the machine mode used for memory references to functions -+being called, in call RTL expressions. On most machines this -+should be QImode. -+*/ -+#define FUNCTION_MODE SImode -+ -+ -+#define REG_S_P(x) \ -+ (REG_P (x) || (GET_CODE (x) == SUBREG && REG_P (XEXP (x, 0)))) -+ -+ -+/* If defined, modifies the length assigned to instruction INSN as a -+ function of the context in which it is used. LENGTH is an lvalue -+ that contains the initially computed length of the insn and should -+ be updated with the correct length of the insn. */ -+#define ADJUST_INSN_LENGTH(INSN, LENGTH) \ -+ ((LENGTH) = avr32_adjust_insn_length ((INSN), (LENGTH))) -+ -+ -+#define CLZ_DEFINED_VALUE_AT_ZERO(mode, value) \ -+ (value = 32, (mode == SImode)) -+ -+#define CTZ_DEFINED_VALUE_AT_ZERO(mode, value) \ -+ (value = 32, (mode == SImode)) -+ -+#define UNITS_PER_SIMD_WORD UNITS_PER_WORD -+ -+#define STORE_FLAG_VALUE 1 -+ -+enum avr32_builtins -+{ -+ AVR32_BUILTIN_MTSR, -+ AVR32_BUILTIN_MFSR, -+ AVR32_BUILTIN_MTDR, -+ AVR32_BUILTIN_MFDR, -+ AVR32_BUILTIN_CACHE, -+ AVR32_BUILTIN_SYNC, -+ AVR32_BUILTIN_TLBR, -+ AVR32_BUILTIN_TLBS, -+ AVR32_BUILTIN_TLBW, -+ AVR32_BUILTIN_BREAKPOINT, -+ AVR32_BUILTIN_XCHG, -+ AVR32_BUILTIN_LDXI, -+ AVR32_BUILTIN_BSWAP16, -+ AVR32_BUILTIN_BSWAP32, -+ AVR32_BUILTIN_COP, -+ AVR32_BUILTIN_MVCR_W, -+ AVR32_BUILTIN_MVRC_W, -+ AVR32_BUILTIN_MVCR_D, -+ AVR32_BUILTIN_MVRC_D, -+ AVR32_BUILTIN_MULSATHH_H, -+ AVR32_BUILTIN_MULSATHH_W, -+ AVR32_BUILTIN_MULSATRNDHH_H, -+ AVR32_BUILTIN_MULSATRNDWH_W, -+ AVR32_BUILTIN_MULSATWH_W, -+ AVR32_BUILTIN_MACSATHH_W, -+ AVR32_BUILTIN_SATADD_H, -+ AVR32_BUILTIN_SATSUB_H, -+ AVR32_BUILTIN_SATADD_W, -+ AVR32_BUILTIN_SATSUB_W, -+ AVR32_BUILTIN_MULWH_D, -+ AVR32_BUILTIN_MULNWH_D, -+ AVR32_BUILTIN_MACWH_D, -+ AVR32_BUILTIN_MACHH_D, -+ AVR32_BUILTIN_MUSFR, -+ AVR32_BUILTIN_MUSTR, -+ AVR32_BUILTIN_SATS, -+ AVR32_BUILTIN_SATU, -+ AVR32_BUILTIN_SATRNDS, -+ AVR32_BUILTIN_SATRNDU -+}; -+ -+ -+#define FLOAT_LIB_COMPARE_RETURNS_BOOL(MODE, COMPARISON) \ -+ ((MODE == SFmode) || (MODE == DFmode)) -+ -+#define RENAME_LIBRARY_SET ".set" -+ -+/* Make ABI_NAME an alias for __GCC_NAME. */ -+#define RENAME_LIBRARY(GCC_NAME, ABI_NAME) \ -+ __asm__ (".globl\t__avr32_" #ABI_NAME "\n" \ -+ ".set\t__avr32_" #ABI_NAME \ -+ ", __" #GCC_NAME "\n"); -+ -+/* Give libgcc functions avr32 ABI name. */ -+#ifdef L_muldi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (muldi3, mul64) -+#endif -+#ifdef L_divdi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (divdi3, sdiv64) -+#endif -+#ifdef L_udivdi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (udivdi3, udiv64) -+#endif -+#ifdef L_moddi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (moddi3, smod64) -+#endif -+#ifdef L_umoddi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (umoddi3, umod64) -+#endif -+#ifdef L_ashldi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (ashldi3, lsl64) -+#endif -+#ifdef L_lshrdi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (lshrdi3, lsr64) -+#endif -+#ifdef L_ashrdi3 -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (ashrdi3, asr64) -+#endif -+ -+#ifdef L_fixsfdi -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (fixsfdi, f32_to_s64) -+#endif -+#ifdef L_fixunssfdi -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (fixunssfdi, f32_to_u64) -+#endif -+#ifdef L_floatdidf -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (floatdidf, s64_to_f64) -+#endif -+#ifdef L_floatdisf -+#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (floatdisf, s64_to_f32) -+#endif -+ -+#endif -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.md ---- gcc-4.2.1/gcc/config/avr32/avr32.md 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.md 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,4606 @@ -+;; AVR32 machine description file. -+;; Copyright 2003-2006 Atmel Corporation. -+;; -+;; Written by Ronny Pedersen, Atmel Norway, -+;; -+;; This file is part of GCC. -+;; -+;; This program is free software; you can redistribute it and/or modify -+;; it under the terms of the GNU General Public License as published by -+;; the Free Software Foundation; either version 2 of the License, or -+;; (at your option) any later version. -+;; -+;; This program is distributed in the hope that it will be useful, -+;; but WITHOUT ANY WARRANTY; without even the implied warranty of -+;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+;; GNU General Public License for more details. -+;; -+;; You should have received a copy of the GNU General Public License -+;; along with this program; if not, write to the Free Software -+;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. -+ -+;; -*- Mode: Scheme -*- -+ -+(define_attr "type" "alu,alu2,alu_sat,mulhh,mulwh,mulww_w,mulww_d,div,machh_w,macww_w,macww_d,branch,call,load,load_rm,store,load2,load4,store2,store4,fmul,fcmps,fcmpd,fcast,fmv,fmvcpu,fldd,fstd,flds,fsts,fstm" -+ (const_string "alu")) -+ -+ -+(define_attr "cc" "none,set_vncz,set_ncz,set_cz,set_z,bld,compare,cmp_cond_insn,clobber,call_set,fpcompare,from_fpcc" -+ (const_string "none")) -+ -+ -+; NB! Keep this in sync with enum architecture_type in avr32.h -+(define_attr "pipeline" "ap,uc" -+ (const (symbol_ref "avr32_arch->arch_type"))) -+ -+(define_attr "length" "" -+ (const_int 4)) -+ -+ -+;; Uses of UNSPEC in this file: -+(define_constants -+ [(UNSPEC_PUSHM 0) -+ (UNSPEC_POPM 1) -+ (UNSPEC_UDIVMODSI4_INTERNAL 2) -+ (UNSPEC_DIVMODSI4_INTERNAL 3) -+ (UNSPEC_STM 4) -+ (UNSPEC_LDM 5) -+ (UNSPEC_MOVSICC 6) -+ (UNSPEC_ADDSICC 7) -+ (UNSPEC_COND_MI 8) -+ (UNSPEC_COND_PL 9) -+ (UNSPEC_PIC_SYM 10) -+ (UNSPEC_PIC_BASE 11) -+ (UNSPEC_STORE_MULTIPLE 12) -+ (UNSPEC_STMFP 13) -+ (UNSPEC_FPCC_TO_REG 14) -+ (UNSPEC_REG_TO_CC 15) -+ (UNSPEC_FORCE_MINIPOOL 16) -+ (UNSPEC_SATS 17) -+ (UNSPEC_SATU 18) -+ (UNSPEC_SATRNDS 19) -+ (UNSPEC_SATRNDU 20) -+ ]) -+ -+(define_constants -+ [(VUNSPEC_EPILOGUE 0) -+ (VUNSPEC_CACHE 1) -+ (VUNSPEC_MTSR 2) -+ (VUNSPEC_MFSR 3) -+ (VUNSPEC_BLOCKAGE 4) -+ (VUNSPEC_SYNC 5) -+ (VUNSPEC_TLBR 6) -+ (VUNSPEC_TLBW 7) -+ (VUNSPEC_TLBS 8) -+ (VUNSPEC_BREAKPOINT 9) -+ (VUNSPEC_MTDR 10) -+ (VUNSPEC_MFDR 11) -+ (VUNSPEC_MVCR 12) -+ (VUNSPEC_MVRC 13) -+ (VUNSPEC_COP 14) -+ (VUNSPEC_ALIGN 15) -+ (VUNSPEC_POOL_START 16) -+ (VUNSPEC_POOL_END 17) -+ (VUNSPEC_POOL_4 18) -+ (VUNSPEC_POOL_8 19) -+ (VUNSPEC_POOL_16 20) -+ (VUNSPEC_MUSFR 21) -+ (VUNSPEC_MUSTR 22) -+ (VUNSPEC_SYNC_CMPXCHG 23) -+ (VUNSPEC_SYNC_SET_LOCK_AND_LOAD 24) -+ (VUNSPEC_SYNC_STORE_IF_LOCK 25) -+ (VUNSPEC_EH_RETURN 26) -+ (VUNSPEC_FRS 27) -+ ]) -+ -+(define_constants -+ [ -+ ;; R7 = 15-7 = 8 -+ (FP_REGNUM 8) -+ ;; Return Register = R12 = 15 - 12 = 3 -+ (RETVAL_REGNUM 3) -+ ;; SP = R13 = 15 - 13 = 2 -+ (SP_REGNUM 2) -+ ;; LR = R14 = 15 - 14 = 1 -+ (LR_REGNUM 1) -+ ;; PC = R15 = 15 - 15 = 0 -+ (PC_REGNUM 0) -+ ;; FPSR = GENERAL_REGS + 1 = 17 -+ (FPCC_REGNUM 17) -+ ]) -+ -+ -+ -+ -+;;****************************************************************************** -+;; Macros -+;;****************************************************************************** -+ -+;; Integer Modes for basic alu insns -+(define_mode_macro INTM [SI HI QI]) -+(define_mode_attr alu_cc_attr [(SI "set_vncz") (HI "clobber") (QI "clobber")]) -+ -+;; Move word modes -+(define_mode_macro MOVM [SI V2HI V4QI]) -+ -+;; For mov/addcc insns -+(define_mode_macro ADDCC [SI HI QI]) -+(define_mode_macro MOVCC [SI HI QI]) -+(define_mode_macro CMP [DI SI HI QI]) -+(define_mode_attr cmp_constraint [(DI "rKu20") (SI "rKs21") (HI "r") (QI "r")]) -+(define_mode_attr cmp_predicate [(DI "register_immediate_operand") -+ (SI "register_immediate_operand") -+ (HI "register_operand") -+ (QI "register_operand")]) -+ -+;; For all conditional insns -+(define_code_macro any_cond [eq ne gt ge lt le gtu geu ltu leu]) -+(define_code_attr cond [(eq "eq") (ne "ne") (gt "gt") (ge "ge") (lt "lt") (le "le") -+ (gtu "hi") (geu "hs") (ltu "lo") (leu "ls")]) -+(define_code_attr invcond [(eq "ne") (ne "eq") (gt "le") (ge "lt") (lt "ge") (le "gt") -+ (gtu "ls") (geu "lo") (ltu "hs") (leu "hi")]) -+ -+;; For logical operations -+(define_code_macro logical [and ior xor]) -+(define_code_attr logical_insn [(and "and") (ior "or") (xor "eor")]) -+ -+;; Load the predicates -+(include "predicates.md") -+ -+ -+;;****************************************************************************** -+;; Automaton pipeline description for avr32 -+;;****************************************************************************** -+ -+(define_automaton "avr32_ap") -+ -+ -+(define_cpu_unit "is" "avr32_ap") -+(define_cpu_unit "a1,m1,da" "avr32_ap") -+(define_cpu_unit "a2,m2,d" "avr32_ap") -+ -+;;Alu instructions -+(define_insn_reservation "alu_op" 1 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "alu")) -+ "is,a1,a2") -+ -+(define_insn_reservation "alu2_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "alu2")) -+ "is,is+a1,a1+a2,a2") -+ -+(define_insn_reservation "alu_sat_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "alu_sat")) -+ "is,a1,a2") -+ -+ -+;;Mul instructions -+(define_insn_reservation "mulhh_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "mulhh,mulwh")) -+ "is,m1,m2") -+ -+(define_insn_reservation "mulww_w_op" 3 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "mulww_w")) -+ "is,m1,m1+m2,m2") -+ -+(define_insn_reservation "mulww_d_op" 5 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "mulww_d")) -+ "is,m1,m1+m2,m1+m2,m2,m2") -+ -+(define_insn_reservation "div_op" 33 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "div")) -+ "is,m1,m1*31 + m2*31,m2") -+ -+(define_insn_reservation "machh_w_op" 3 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "machh_w")) -+ "is*2,m1,m2") -+ -+ -+(define_insn_reservation "macww_w_op" 4 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "macww_w")) -+ "is*2,m1,m1,m2") -+ -+ -+(define_insn_reservation "macww_d_op" 6 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "macww_d")) -+ "is*2,m1,m1+m2,m1+m2,m2") -+ -+;;Bypasses for Mac instructions, because of accumulator cache. -+;;Set latency as low as possible in order to let the compiler let -+;;mul -> mac and mac -> mac combinations which use the same -+;;accumulator cache be placed close together to avoid any -+;;instructions which can ruin the accumulator cache come inbetween. -+(define_bypass 4 "machh_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+(define_bypass 5 "macww_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+(define_bypass 7 "macww_d_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+ -+(define_bypass 3 "mulhh_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+(define_bypass 4 "mulww_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+(define_bypass 6 "mulww_d_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") -+ -+ -+;;Bypasses for all mul/mac instructions followed by an instruction -+;;which reads the output AND writes the result to the same register. -+;;This will generate an Write After Write hazard which gives an -+;;extra cycle before the result is ready. -+(define_bypass 0 "machh_w_op" "machh_w_op" "avr32_valid_macmac_bypass") -+(define_bypass 0 "macww_w_op" "macww_w_op" "avr32_valid_macmac_bypass") -+(define_bypass 0 "macww_d_op" "macww_d_op" "avr32_valid_macmac_bypass") -+ -+(define_bypass 0 "mulhh_op" "machh_w_op" "avr32_valid_mulmac_bypass") -+(define_bypass 0 "mulww_w_op" "macww_w_op" "avr32_valid_mulmac_bypass") -+(define_bypass 0 "mulww_d_op" "macww_d_op" "avr32_valid_mulmac_bypass") -+ -+;;Branch and call instructions -+;;We assume that all branches and rcalls are predicted correctly :-) -+;;while calls use a lot of cycles. -+(define_insn_reservation "branch_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "branch")) -+ "nothing") -+ -+(define_insn_reservation "call_op" 10 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "call")) -+ "nothing") -+ -+ -+;;Load store instructions -+(define_insn_reservation "load_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "load")) -+ "is,da,d") -+ -+(define_insn_reservation "load_rm_op" 3 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "load_rm")) -+ "is,da,d") -+ -+ -+(define_insn_reservation "store_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "store")) -+ "is,da,d") -+ -+ -+(define_insn_reservation "load_double_op" 3 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "load2")) -+ "is,da,da+d,d") -+ -+(define_insn_reservation "load_quad_op" 4 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "load4")) -+ "is,da,da+d,da+d,d") -+ -+(define_insn_reservation "store_double_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "store2")) -+ "is,da,da+d,d") -+ -+ -+(define_insn_reservation "store_quad_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "store4")) -+ "is,da,da+d,da+d,d") -+ -+;;For store the operand to write to memory is read in d and -+;;the real latency between any instruction and a store is therefore -+;;one less than for the instructions which reads the operands in the first -+;;excecution stage -+(define_bypass 2 "load_double_op" "store_double_op" "avr32_store_bypass") -+(define_bypass 3 "load_quad_op" "store_quad_op" "avr32_store_bypass") -+(define_bypass 1 "load_op" "store_op" "avr32_store_bypass") -+(define_bypass 2 "load_rm_op" "store_op" "avr32_store_bypass") -+(define_bypass 1 "alu_sat_op" "store_op" "avr32_store_bypass") -+(define_bypass 1 "alu2_op" "store_op" "avr32_store_bypass") -+(define_bypass 1 "mulhh_op" "store_op" "avr32_store_bypass") -+(define_bypass 2 "mulww_w_op" "store_op" "avr32_store_bypass") -+(define_bypass 4 "mulww_d_op" "store_op" "avr32_store_bypass" ) -+(define_bypass 2 "machh_w_op" "store_op" "avr32_store_bypass") -+(define_bypass 3 "macww_w_op" "store_op" "avr32_store_bypass") -+(define_bypass 5 "macww_d_op" "store_op" "avr32_store_bypass") -+ -+ -+; Bypass for load double operation. If only the first loaded word is needed -+; then the latency is 2 -+(define_bypass 2 "load_double_op" -+ "load_op,load_rm_op,alu_sat_op, alu2_op, alu_op, mulhh_op, mulww_w_op, -+ mulww_d_op, machh_w_op, macww_w_op, macww_d_op" -+ "avr32_valid_load_double_bypass") -+ -+; Bypass for load quad operation. If only the first or second loaded word is needed -+; we set the latency to 2 -+(define_bypass 2 "load_quad_op" -+ "load_op,load_rm_op,alu_sat_op, alu2_op, alu_op, mulhh_op, mulww_w_op, -+ mulww_d_op, machh_w_op, macww_w_op, macww_d_op" -+ "avr32_valid_load_quad_bypass") -+ -+ -+;;****************************************************************************** -+;; End of Automaton pipeline description for avr32 -+;;****************************************************************************** -+ -+ -+ -+;;============================================================================= -+;; move -+;;----------------------------------------------------------------------------- -+ -+;;== char - 8 bits ============================================================ -+(define_expand "movqi" -+ [(set (match_operand:QI 0 "nonimmediate_operand" "") -+ (match_operand:QI 1 "general_operand" ""))] -+ "" -+ { -+ if ( !no_new_pseudos ){ -+ if (GET_CODE (operands[1]) == MEM && optimize){ -+ rtx reg = gen_reg_rtx (SImode); -+ -+ emit_insn (gen_zero_extendqisi2 (reg, operands[1])); -+ operands[1] = gen_lowpart (QImode, reg); -+ } -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) == MEM) -+ operands[1] = force_reg (QImode, operands[1]); -+ } -+ -+ }) -+ -+(define_insn "*movqi_internal" -+ [(set (match_operand:QI 0 "nonimmediate_operand" "=r,r,m,r") -+ (match_operand:QI 1 "general_operand" "rKs08,m,r,i"))] -+ "" -+ "@ -+ mov\t%0, %1 -+ ld.ub\t%0, %1 -+ st.b\t%0, %1 -+ mov\t%0, %1" -+ [(set_attr "length" "2,4,4,4") -+ (set_attr "type" "alu,load_rm,store,alu")]) -+ -+ -+ -+;;== short - 16 bits ========================================================== -+(define_expand "movhi" -+ [(set (match_operand:HI 0 "nonimmediate_operand" "") -+ (match_operand:HI 1 "general_operand" ""))] -+ "" -+ { -+ if ( !no_new_pseudos ){ -+ if (GET_CODE (operands[1]) == MEM && optimize){ -+ rtx reg = gen_reg_rtx (SImode); -+ -+ emit_insn (gen_extendhisi2 (reg, operands[1])); -+ operands[1] = gen_lowpart (HImode, reg); -+ } -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) == MEM) -+ operands[1] = force_reg (HImode, operands[1]); -+ } -+ -+ }) -+ -+(define_insn "*movhi_internal" -+ [(set (match_operand:HI 0 "nonimmediate_operand" "=r,r,m,r") -+ (match_operand:HI 1 "general_operand" "rKs08,m,r,i"))] -+ "" -+ "@ -+ mov\t%0, %1 -+ ld.sh\t%0, %1 -+ st.h\t%0, %1 -+ mov\t%0, %1" -+ [(set_attr "length" "2,4,4,4") -+ (set_attr "type" "alu,load_rm,store,alu")]) -+ -+ -+;;== int - 32 bits ============================================================ -+ -+(define_expand "movmisalignsi" -+ [(set (match_operand:SI 0 "nonimmediate_operand" "") -+ (match_operand:SI 1 "nonimmediate_operand" ""))] -+ "TARGET_UNALIGNED_WORD" -+ { -+ } -+) -+ -+(define_expand "mov" -+ [(set (match_operand:MOVM 0 "register_operand" "") -+ (match_operand:MOVM 1 "general_operand" ""))] -+ "" -+ { -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) == MEM) -+ operands[1] = force_reg (mode, operands[1]); -+ -+ -+ /* Check for out of range immediate constants as these may -+ occur during reloading, since it seems like reload does -+ not check if the immediate is legitimate. Don't know if -+ this is a bug? */ -+ if ( reload_in_progress -+ && avr32_arch->arch_type != ARCH_TYPE_AVR32_UC -+ && GET_CODE(operands[1]) == CONST_INT -+ && !avr32_const_ok_for_constraint_p(INTVAL(operands[1]), 'K', "Ks21") ){ -+ operands[1] = force_const_mem(SImode, operands[1]); -+ } -+ -+ if ( (flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS) -+ && !avr32_legitimate_pic_operand_p(operands[1]) ) -+ operands[1] = legitimize_pic_address (operands[1], mode, -+ (no_new_pseudos ? operands[0] : 0)); -+ else if ( flag_pic && avr32_address_operand(operands[1], GET_MODE(operands[1])) ) -+ /* If we have an address operand then this function uses the pic register. */ -+ current_function_uses_pic_offset_table = 1; -+ }) -+ -+ -+(define_insn "mov_internal" -+ [(set (match_operand:MOVM 0 "nonimmediate_operand" "=r, r, r,r,m,r") -+ (match_operand:MOVM 1 "general_operand" "rKs08,Ks21,n,m,r,W"))] -+ "" -+ { -+ switch (which_alternative) { -+ case 0: -+ case 1: return "mov\t%0, %1"; -+ case 2: return "mov\t%0, lo(%1)\;orh\t%0,hi(%1)"; -+ case 3: -+ if ( (REG_P(XEXP(operands[1], 0)) -+ && REGNO(XEXP(operands[1], 0)) == SP_REGNUM) -+ || (GET_CODE(XEXP(operands[1], 0)) == PLUS -+ && REGNO(XEXP(XEXP(operands[1], 0), 0)) == SP_REGNUM -+ && GET_CODE(XEXP(XEXP(operands[1], 0), 1)) == CONST_INT -+ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) % 4 == 0 -+ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) <= 0x1FC) ) -+ return "lddsp\t%0, %1"; -+ else if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1])) ) -+ return "lddpc\t%0, %1"; -+ else -+ return "ld.w\t%0, %1"; -+ case 4: -+ if ( (REG_P(XEXP(operands[0], 0)) -+ && REGNO(XEXP(operands[0], 0)) == SP_REGNUM) -+ || (GET_CODE(XEXP(operands[0], 0)) == PLUS -+ && REGNO(XEXP(XEXP(operands[0], 0), 0)) == SP_REGNUM -+ && GET_CODE(XEXP(XEXP(operands[0], 0), 1)) == CONST_INT -+ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) % 4 == 0 -+ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) <= 0x1FC) ) -+ return "stdsp\t%0, %1"; -+ else -+ return "st.w\t%0, %1"; -+ case 5: -+ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) -+ return "lda.w\t%0, %1"; -+ else -+ return "ld.w\t%0, r6[%1@got]"; -+ default: -+ abort(); -+ } -+ } -+ -+ [(set_attr "length" "2,4,8,4,4,8") -+ (set_attr "type" "alu,alu,alu2,load,store,load") -+ (set_attr "cc" "none,none,set_z,none,none,clobber")]) -+ -+ -+;; These instructions are for loading constants which cannot be loaded -+;; directly from the constant pool because the offset is too large -+;; high and lo_sum are used even tough for our case it should be -+;; low and high sum :-) -+(define_insn "mov_symbol_lo" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (high:SI (match_operand:SI 1 "immediate_operand" "i" )))] -+ "" -+ "mov\t%0, lo(%1)" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4")] -+) -+ -+(define_insn "add_symbol_hi" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (lo_sum:SI (match_dup 0) -+ (match_operand:SI 1 "immediate_operand" "i" )))] -+ "" -+ "orh\t%0, hi(%1)" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4")] -+) -+ -+ -+ -+;; When generating pic, we need to load the symbol offset into a register. -+;; So that the optimizer does not confuse this with a normal symbol load -+;; we use an unspec. The offset will be loaded from a constant pool entry, -+;; since that is the only type of relocation we can use. -+(define_insn "pic_load_addr" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (unspec:SI [(match_operand:SI 1 "" "")] UNSPEC_PIC_SYM))] -+ "flag_pic && CONSTANT_POOL_ADDRESS_P(XEXP(operands[1], 0))" -+ "lddpc\t%0, %1" -+ [(set_attr "type" "load") -+ (set_attr "length" "4")] -+) -+ -+(define_insn "pic_compute_got_from_pc" -+ [(set (match_operand:SI 0 "register_operand" "+r") -+ (unspec:SI [(minus:SI (pc) -+ (match_dup 0))] UNSPEC_PIC_BASE)) -+ (use (label_ref (match_operand 1 "" "")))] -+ "flag_pic" -+ { -+ (*targetm.asm_out.internal_label) (asm_out_file, "L", -+ CODE_LABEL_NUMBER (operands[1])); -+ return \"rsub\t%0, pc\"; -+ } -+ [(set_attr "cc" "clobber") -+ (set_attr "length" "2")] -+) -+ -+;;== long long int - 64 bits ================================================== -+ -+(define_expand "movdi" -+ [(set (match_operand:DI 0 "nonimmediate_operand" "") -+ (match_operand:DI 1 "general_operand" ""))] -+ "" -+ { -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) != REG) -+ operands[1] = force_reg (DImode, operands[1]); -+ -+ }) -+ -+ -+(define_insn_and_split "*movdi_internal" -+ [(set (match_operand:DI 0 "nonimmediate_operand" "=r,r, r, r,r,r,m") -+ (match_operand:DI 1 "general_operand" "r, Ks08,Ks21,G,n,m,r"))] -+ "" -+ { -+ switch (which_alternative ){ -+ case 0: -+ case 1: -+ case 2: -+ case 3: -+ case 4: -+ return "#"; -+ case 5: -+ if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1]))) -+ return "ld.d\t%0, pc[%1 - .]"; -+ else -+ return "ld.d\t%0, %1"; -+ case 6: -+ return "st.d\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+;; Lets split all reg->reg or imm->reg transfers into two SImode transfers -+ "reload_completed && -+ (REG_P (operands[0]) && -+ (REG_P (operands[1]) -+ || GET_CODE (operands[1]) == CONST_INT -+ || GET_CODE (operands[1]) == CONST_DOUBLE))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 2) (match_dup 3))] -+ { -+ operands[2] = gen_highpart (SImode, operands[0]); -+ operands[0] = gen_lowpart (SImode, operands[0]); -+ if ( REG_P(operands[1]) ){ -+ operands[3] = gen_highpart(SImode, operands[1]); -+ operands[1] = gen_lowpart(SImode, operands[1]); -+ } else if ( GET_CODE(operands[1]) == CONST_DOUBLE -+ || GET_CODE(operands[1]) == CONST_INT ){ -+ rtx split_const[2]; -+ avr32_split_const_expr (DImode, SImode, operands[1], split_const); -+ operands[3] = split_const[1]; -+ operands[1] = split_const[0]; -+ } else { -+ internal_error("Illegal operand[1] for movdi split!"); -+ } -+ } -+ -+ [(set_attr "length" "*,*,*,*,*,4,4") -+ (set_attr "type" "*,*,*,*,*,load2,store2") -+ (set_attr "cc" "*,*,*,*,*,none,none")]) -+ -+ -+;;== 128 bits ================================================== -+(define_expand "movti" -+ [(set (match_operand:TI 0 "nonimmediate_operand" "") -+ (match_operand:TI 1 "nonimmediate_operand" ""))] -+ "" -+ { -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) != REG) -+ operands[1] = force_reg (TImode, operands[1]); -+ -+ /* We must fix any pre_dec for loads and post_inc stores */ -+ if ( GET_CODE (operands[0]) == MEM -+ && GET_CODE (XEXP(operands[0],0)) == POST_INC ){ -+ emit_move_insn(gen_rtx_MEM(TImode, XEXP(XEXP(operands[0],0),0)), operands[1]); -+ emit_insn(gen_addsi3(XEXP(XEXP(operands[0],0),0), XEXP(XEXP(operands[0],0),0), GEN_INT(GET_MODE_SIZE(TImode)))); -+ DONE; -+ } -+ -+ if ( GET_CODE (operands[1]) == MEM -+ && GET_CODE (XEXP(operands[1],0)) == PRE_DEC ){ -+ emit_insn(gen_addsi3(XEXP(XEXP(operands[1],0),0), XEXP(XEXP(operands[1],0),0), GEN_INT(-GET_MODE_SIZE(TImode)))); -+ emit_move_insn(operands[0], gen_rtx_MEM(TImode, XEXP(XEXP(operands[1],0),0))); -+ DONE; -+ } -+ }) -+ -+ -+(define_insn_and_split "*movti_internal" -+ [(set (match_operand:TI 0 "avr32_movti_dst_operand" "=r,r,,r,n,m"))] -+ "" -+ { -+ switch (which_alternative ){ -+ case 0: -+ case 3: -+ return "#"; -+ case 1: -+ return "ldm\t%p1, %0"; -+ case 2: -+ return "stm\t%p0, %1"; -+ case 4: -+ return "ld.d\t%U0, pc[%1 - .]\;ld.d\t%B0, pc[%1 - . + 8]"; -+ } -+ } -+ -+ "reload_completed && -+ (REG_P (operands[0]) && -+ (REG_P (operands[1]) -+ || (GET_CODE (operands[1]) == MEM -+ && GET_CODE (XEXP (operands[1], 0)) == SYMBOL_REF -+ && CONSTANT_POOL_ADDRESS_P (XEXP (operands[1], 0))) -+ || GET_CODE (operands[1]) == CONST_INT -+ || GET_CODE (operands[1]) == CONST_DOUBLE))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 2) (match_dup 3))] -+ { -+ operands[2] = simplify_gen_subreg ( DImode, operands[0], -+ TImode, 0 ); -+ operands[0] = simplify_gen_subreg ( DImode, operands[0], -+ TImode, 8 ); -+ if ( REG_P(operands[1]) ){ -+ operands[3] = simplify_gen_subreg ( DImode, operands[1], -+ TImode, 0 ); -+ operands[1] = simplify_gen_subreg ( DImode, operands[1], -+ TImode, 8 ); -+ } else if ( GET_CODE(operands[1]) == CONST_DOUBLE -+ || GET_CODE(operands[1]) == CONST_INT ){ -+ rtx split_const[2]; -+ avr32_split_const_expr (TImode, DImode, operands[1], split_const); -+ operands[3] = split_const[1]; -+ operands[1] = split_const[0]; -+ } else if (avr32_const_pool_ref_operand (operands[1], GET_MODE(operands[1]))){ -+ rtx split_const[2]; -+ rtx cop = avoid_constant_pool_reference (operands[1]); -+ if (operands[1] == cop) -+ cop = get_pool_constant (XEXP (operands[1], 0)); -+ avr32_split_const_expr (TImode, DImode, cop, split_const); -+ operands[3] = force_const_mem (DImode, split_const[1]); -+ operands[1] = force_const_mem (DImode, split_const[0]); -+ } else { -+ internal_error("Illegal operand[1] for movdi split!"); -+ } -+ } -+ [(set_attr "length" "*,4,4,*,8") -+ (set_attr "type" "*,load4,store4,*,load4")]) -+ -+ -+;;== float - 32 bits ========================================================== -+(define_expand "movsf" -+ [(set (match_operand:SF 0 "nonimmediate_operand" "") -+ (match_operand:SF 1 "general_operand" ""))] -+ "" -+ { -+ -+ -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) != REG) -+ operands[1] = force_reg (SFmode, operands[1]); -+ -+ }) -+ -+(define_insn "*movsf_internal" -+ [(set (match_operand:SF 0 "nonimmediate_operand" "=r,r,r,r,m") -+ (match_operand:SF 1 "general_operand" "r, G,F,m,r"))] -+ "TARGET_SOFT_FLOAT" -+ { -+ switch (which_alternative) { -+ case 0: -+ case 1: return "mov\t%0, %1"; -+ case 2: return "mov\t%0, lo(%1)\;orh\t%0, hi(%1)"; -+ case 3: -+ if ( (REG_P(XEXP(operands[1], 0)) -+ && REGNO(XEXP(operands[1], 0)) == SP_REGNUM) -+ || (GET_CODE(XEXP(operands[1], 0)) == PLUS -+ && REGNO(XEXP(XEXP(operands[1], 0), 0)) == SP_REGNUM -+ && GET_CODE(XEXP(XEXP(operands[1], 0), 1)) == CONST_INT -+ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) % 4 == 0 -+ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) <= 0x1FC) ) -+ return "lddsp\t%0, %1"; -+ else if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1])) ) -+ return "lddpc\t%0, %1"; -+ else -+ return "ld.w\t%0, %1"; -+ case 4: -+ if ( (REG_P(XEXP(operands[0], 0)) -+ && REGNO(XEXP(operands[0], 0)) == SP_REGNUM) -+ || (GET_CODE(XEXP(operands[0], 0)) == PLUS -+ && REGNO(XEXP(XEXP(operands[0], 0), 0)) == SP_REGNUM -+ && GET_CODE(XEXP(XEXP(operands[0], 0), 1)) == CONST_INT -+ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) % 4 == 0 -+ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) <= 0x1FC) ) -+ return "stdsp\t%0, %1"; -+ else -+ return "st.w\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ -+ [(set_attr "length" "2,4,8,4,4") -+ (set_attr "type" "alu,alu,alu2,load,store") -+ (set_attr "cc" "none,none,clobber,none,none")]) -+ -+ -+ -+;;== double - 64 bits ========================================================= -+(define_expand "movdf" -+ [(set (match_operand:DF 0 "nonimmediate_operand" "") -+ (match_operand:DF 1 "general_operand" ""))] -+ "" -+ { -+ /* One of the ops has to be in a register. */ -+ if (GET_CODE (operands[0]) != REG){ -+ operands[1] = force_reg (DFmode, operands[1]); -+ } -+ }) -+ -+ -+(define_insn_and_split "*movdf_internal" -+ [(set (match_operand:DF 0 "nonimmediate_operand" "=r,r,r,r,m") -+ (match_operand:DF 1 "general_operand" " r,G,F,m,r"))] -+ "TARGET_SOFT_FLOAT" -+ { -+ switch (which_alternative ){ -+ case 0: -+ case 1: -+ case 2: -+ return "#"; -+ case 3: -+ if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1]))) -+ return "ld.d\t%0, pc[%1 - .]"; -+ else -+ return "ld.d\t%0, %1"; -+ case 4: -+ return "st.d\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ "TARGET_SOFT_FLOAT -+ && reload_completed -+ && (REG_P (operands[0]) -+ && (REG_P (operands[1]) -+ || GET_CODE (operands[1]) == CONST_DOUBLE))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 2) (match_dup 3))] -+ " -+ { -+ operands[2] = gen_highpart (SImode, operands[0]); -+ operands[0] = gen_lowpart (SImode, operands[0]); -+ operands[3] = gen_highpart(SImode, operands[1]); -+ operands[1] = gen_lowpart(SImode, operands[1]); -+ } -+ " -+ -+ [(set_attr "length" "*,*,*,4,4") -+ (set_attr "type" "*,*,*,load2,store2") -+ (set_attr "cc" "*,*,*,none,none")]) -+ -+ -+ -+ -+;;============================================================================= -+;; Move chunks of memory -+;;============================================================================= -+ -+(define_expand "movmemsi" -+ [(match_operand:BLK 0 "general_operand" "") -+ (match_operand:BLK 1 "general_operand" "") -+ (match_operand:SI 2 "const_int_operand" "") -+ (match_operand:SI 3 "const_int_operand" "")] -+ "" -+ " -+ if (avr32_gen_movmemsi (operands)) -+ DONE; -+ FAIL; -+ " -+ -+ ) -+ -+ -+ -+;;============================================================================= -+;; Bit field instructions -+;;----------------------------------------------------------------------------- -+;; Instructions to insert or extract bit-fields -+;;============================================================================= -+ -+(define_insn "insv" -+ [ (set (zero_extract:SI (match_operand:SI 0 "register_operand" "+r") -+ (match_operand:SI 1 "immediate_operand" "Ku05") -+ (match_operand:SI 2 "immediate_operand" "Ku05")) -+ (match_operand 3 "register_operand" "r"))] -+ "" -+ "bfins\t%0, %3, %2, %1" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "set_ncz")]) -+ -+ -+ -+(define_expand "extv" -+ [ (set (match_operand:SI 0 "register_operand" "") -+ (sign_extract:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "") -+ (match_operand:SI 3 "immediate_operand" "")))] -+ "" -+ { -+ if ( INTVAL(operands[2]) >= 32 ) -+ FAIL; -+ } -+) -+ -+(define_expand "extzv" -+ [ (set (match_operand:SI 0 "register_operand" "") -+ (zero_extract:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "") -+ (match_operand:SI 3 "immediate_operand" "")))] -+ "" -+ { -+ if ( INTVAL(operands[2]) >= 32 ) -+ FAIL; -+ } -+) -+ -+(define_insn "extv_internal" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (sign_extract:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku05") -+ (match_operand:SI 3 "immediate_operand" "Ku05")))] -+ "INTVAL(operands[2]) < 32" -+ "bfexts\t%0, %1, %3, %2" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "set_ncz")]) -+ -+ -+(define_insn "extzv_internal" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (zero_extract:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku05") -+ (match_operand:SI 3 "immediate_operand" "Ku05")))] -+ "INTVAL(operands[2]) < 32" -+ "bfextu\t%0, %1, %3, %2" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "set_ncz")]) -+ -+ -+ -+;;============================================================================= -+;; Some peepholes for avoiding unnecessary cast instructions -+;; followed by bfins. -+;;----------------------------------------------------------------------------- -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (zero_extend:SI (match_operand:QI 1 "register_operand" ""))) -+ (set (zero_extract:SI (match_operand 2 "register_operand" "") -+ (match_operand:SI 3 "immediate_operand" "") -+ (match_operand:SI 4 "immediate_operand" "")) -+ (match_dup 0))] -+ "((peep2_reg_dead_p(2, operands[0]) && -+ (INTVAL(operands[3]) <= 8)))" -+ [(set (zero_extract:SI (match_dup 2) -+ (match_dup 3) -+ (match_dup 4)) -+ (match_dup 1))] -+ ) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (zero_extend:SI (match_operand:HI 1 "register_operand" ""))) -+ (set (zero_extract:SI (match_operand 2 "register_operand" "") -+ (match_operand:SI 3 "immediate_operand" "") -+ (match_operand:SI 4 "immediate_operand" "")) -+ (match_dup 0))] -+ "((peep2_reg_dead_p(2, operands[0]) && -+ (INTVAL(operands[3]) <= 16)))" -+ [(set (zero_extract:SI (match_dup 2) -+ (match_dup 3) -+ (match_dup 4)) -+ (match_dup 1))] -+ ) -+ -+;;============================================================================= -+;; push bytes -+;;----------------------------------------------------------------------------- -+;; Implements the push instruction -+;;============================================================================= -+(define_insn "pushm" -+ [(set (mem:BLK (pre_dec:BLK (reg:SI SP_REGNUM))) -+ (unspec:BLK [(match_operand 0 "const_int_operand" "")] -+ UNSPEC_PUSHM))] -+ "" -+ { -+ if (INTVAL(operands[0])) { -+ return "pushm\t%r0"; -+ } else { -+ return ""; -+ } -+ } -+ [(set_attr "type" "store") -+ (set_attr "length" "2") -+ (set_attr "cc" "none")]) -+ -+(define_insn "stm" -+ [(unspec [(match_operand 0 "register_operand" "r") -+ (match_operand 1 "const_int_operand" "") -+ (match_operand 2 "const_int_operand" "")] -+ UNSPEC_STM)] -+ "" -+ { -+ if (INTVAL(operands[1])) { -+ if (INTVAL(operands[2]) != 0) -+ return "stm\t--%0, %s1"; -+ else -+ return "stm\t%0, %s1"; -+ } else { -+ return ""; -+ } -+ } -+ [(set_attr "type" "store") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+ -+ -+(define_insn "popm" -+ [(unspec [(match_operand 0 "const_int_operand" "")] -+ UNSPEC_POPM)] -+ "" -+ { -+ if (INTVAL(operands[0])) { -+ return "popm %r0"; -+ } else { -+ return ""; -+ } -+ } -+ [(set_attr "type" "load") -+ (set_attr "length" "2")]) -+ -+ -+ -+;;============================================================================= -+;; add -+;;----------------------------------------------------------------------------- -+;; Adds reg1 with reg2 and puts the result in reg0. -+;;============================================================================= -+(define_insn "add3" -+ [(set (match_operand:INTM 0 "register_operand" "=r,r,r,r,r") -+ (plus:INTM (match_operand:INTM 1 "register_operand" "%0,r,0,r,0") -+ (match_operand:INTM 2 "avr32_add_operand" "r,r,Is08,Is16,Is21")))] -+ "" -+ "@ -+ add %0, %2 -+ add %0, %1, %2 -+ sub %0, %n2 -+ sub %0, %1, %n2 -+ sub %0, %n2" -+ -+ [(set_attr "length" "2,4,2,4,4") -+ (set_attr "cc" "")]) -+ -+(define_insn "add3_lsl" -+ [(set (match_operand:INTM 0 "register_operand" "=r") -+ (plus:INTM (ashift:INTM (match_operand:INTM 1 "register_operand" "r") -+ (match_operand:INTM 3 "avr32_add_shift_immediate_operand" "Ku02")) -+ (match_operand:INTM 2 "register_operand" "r")))] -+ "" -+ "add %0, %2, %1 << %3" -+ [(set_attr "length" "4") -+ (set_attr "cc" "")]) -+ -+(define_insn "add3_lsl2" -+ [(set (match_operand:INTM 0 "register_operand" "=r") -+ (plus:INTM (match_operand:INTM 1 "register_operand" "r") -+ (ashift:INTM (match_operand:INTM 2 "register_operand" "r") -+ (match_operand:INTM 3 "avr32_add_shift_immediate_operand" "Ku02"))))] -+ "" -+ "add %0, %1, %2 << %3" -+ [(set_attr "length" "4") -+ (set_attr "cc" "")]) -+ -+ -+ -+(define_insn "add3_mul" -+ [(set (match_operand:INTM 0 "register_operand" "=r") -+ (plus:INTM (mult:INTM (match_operand:INTM 1 "register_operand" "r") -+ (match_operand:INTM 3 "immediate_operand" "Ku04" )) -+ (match_operand:INTM 2 "register_operand" "r")))] -+ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || -+ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" -+ "add %0, %2, %1 << %p3" -+ [(set_attr "length" "4") -+ (set_attr "cc" "")]) -+ -+(define_insn "add3_mul2" -+ [(set (match_operand:INTM 0 "register_operand" "=r") -+ (plus:INTM (match_operand:INTM 1 "register_operand" "r") -+ (mult:INTM (match_operand:INTM 2 "register_operand" "r") -+ (match_operand:INTM 3 "immediate_operand" "Ku04" ))))] -+ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || -+ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" -+ "add %0, %1, %2 << %p3" -+ [(set_attr "length" "4") -+ (set_attr "cc" "")]) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (ashift:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (plus:SI (match_dup 0) -+ (match_operand:SI 4 "register_operand" "")))] -+ "(peep2_reg_dead_p(2, operands[0]) && -+ (INTVAL(operands[2]) < 4 && INTVAL(operands[2]) > 0))" -+ [(set (match_dup 3) -+ (plus:SI (ashift:SI (match_dup 1) -+ (match_dup 2)) -+ (match_dup 4)))] -+ ) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (ashift:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (plus:SI (match_operand:SI 4 "register_operand" "") -+ (match_dup 0)))] -+ "(peep2_reg_dead_p(2, operands[0]) && -+ (INTVAL(operands[2]) < 4 && INTVAL(operands[2]) > 0))" -+ [(set (match_dup 3) -+ (plus:SI (ashift:SI (match_dup 1) -+ (match_dup 2)) -+ (match_dup 4)))] -+ ) -+ -+(define_insn "adddi3" -+ [(set (match_operand:DI 0 "register_operand" "=r,r") -+ (plus:DI (match_operand:DI 1 "register_operand" "%r,0") -+ (match_operand:DI 2 "register_operand" "r,r")))] -+ "" -+ "@ -+ add %0, %1, %2\;adc %m0, %m1, %m2 -+ add %0, %2\;adc %m0, %m0, %m2" -+ [(set_attr "length" "8,6") -+ (set_attr "type" "alu2") -+ (set_attr "cc" "set_vncz")]) -+ -+ -+ -+;;============================================================================= -+;; subtract -+;;----------------------------------------------------------------------------- -+;; Subtract reg2 or immediate value from reg0 and puts the result in reg0. -+;;============================================================================= -+ -+(define_insn "sub3" -+ [(set (match_operand:INTM 0 "general_operand" "=r,r,r,r,r,r,r") -+ (minus:INTM (match_operand:INTM 1 "nonmemory_operand" "0,r,0,r,0,r,Ks08") -+ (match_operand:INTM 2 "nonmemory_operand" "r,r,Ks08,Ks16,Ks21,0,r")))] -+ "" -+ "@ -+ sub %0, %2 -+ sub %0, %1, %2 -+ sub %0, %2 -+ sub %0, %1, %2 -+ sub %0, %2 -+ rsub %0, %1 -+ rsub %0, %2, %1" -+ [(set_attr "length" "2,4,2,4,4,2,4") -+ (set_attr "cc" "")]) -+ -+(define_insn "*sub3_mul" -+ [(set (match_operand:INTM 0 "register_operand" "=r,r,r") -+ (minus:INTM (match_operand:INTM 1 "register_operand" "r,0,r") -+ (mult:INTM (match_operand:INTM 2 "register_operand" "r,r,0") -+ (match_operand:SI 3 "immediate_operand" "Ku04,Ku04,Ku04" ))))] -+ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || -+ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" -+ "@ -+ sub %0, %1, %2 << %p3 -+ sub %0, %0, %2 << %p3 -+ sub %0, %1, %0 << %p3" -+ [(set_attr "length" "4,4,4") -+ (set_attr "cc" "")]) -+ -+(define_insn "*sub3_lsl" -+ [(set (match_operand:INTM 0 "register_operand" "=r") -+ (minus:INTM (match_operand:INTM 1 "register_operand" "r") -+ (ashift:INTM (match_operand:INTM 2 "register_operand" "r") -+ (match_operand:SI 3 "avr32_add_shift_immediate_operand" "Ku02"))))] -+ "" -+ "sub %0, %1, %2 << %3" -+ [(set_attr "length" "4") -+ (set_attr "cc" "")]) -+ -+ -+(define_insn "subdi3" -+ [(set (match_operand:DI 0 "register_operand" "=r,r") -+ (minus:DI (match_operand:DI 1 "register_operand" "%r,0") -+ (match_operand:DI 2 "register_operand" "r,r")))] -+ "" -+ "@ -+ sub %0, %1, %2\;sbc %m0, %m1, %m2 -+ sub %0, %2\;sbc %m0, %m0, %m2" -+ [(set_attr "length" "8,6") -+ (set_attr "type" "alu2") -+ (set_attr "cc" "set_vncz")]) -+ -+ -+ -+;;============================================================================= -+;; multiply -+;;----------------------------------------------------------------------------- -+;; Multiply op1 and op2 and put the value in op0. -+;;============================================================================= -+ -+ -+(define_insn "mulqi3" -+ [(set (match_operand:QI 0 "register_operand" "=r,r,r") -+ (mult:QI (match_operand:QI 1 "register_operand" "%0,r,r") -+ (match_operand:QI 2 "avr32_mul_operand" "r,r,Ks08")))] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "mul %0, %2"; -+ case 1: -+ return "mul %0, %1, %2"; -+ case 2: -+ return "mul %0, %1, %2"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "type" "mulww_w,mulww_w,mulwh") -+ (set_attr "length" "2,4,4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "mulsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (mult:SI (match_operand:SI 1 "register_operand" "%0,r,r") -+ (match_operand:SI 2 "avr32_mul_operand" "r,r,Ks08")))] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "mul %0, %2"; -+ case 1: -+ return "mul %0, %1, %2"; -+ case 2: -+ return "mul %0, %1, %2"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "type" "mulww_w,mulww_w,mulwh") -+ (set_attr "length" "2,4,4") -+ (set_attr "cc" "none")]) -+ -+ -+(define_insn "mulhisi3" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (mult:SI -+ (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))))] -+ "TARGET_DSP" -+ "mulhh.w %0, %1:b, %2:b" -+ [(set_attr "type" "mulhh") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_peephole2 -+ [(match_scratch:DI 6 "r") -+ (set (match_operand:SI 0 "register_operand" "") -+ (mult:SI -+ (sign_extend:SI (match_operand:HI 1 "register_operand" "")) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "")))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (ashiftrt:SI (match_dup 0) -+ (const_int 16)))] -+ "TARGET_DSP -+ && (peep2_reg_dead_p(1, operands[0]) || (REGNO(operands[0]) == REGNO(operands[3])))" -+ [(set (match_dup 4) (sign_extend:SI (match_dup 1))) -+ (set (match_dup 6) -+ (ashift:DI (mult:DI (sign_extend:DI (match_dup 4)) -+ (sign_extend:DI (match_dup 2))) -+ (const_int 16))) -+ (set (match_dup 3) (match_dup 5))] -+ -+ "{ -+ operands[4] = gen_rtx_REG(SImode, REGNO(operands[1])); -+ operands[5] = gen_highpart (SImode, operands[4]); -+ }" -+ ) -+ -+(define_insn "mulnhisi3" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (mult:SI -+ (sign_extend:SI (neg:HI (match_operand:HI 1 "register_operand" "r"))) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))))] -+ "TARGET_DSP" -+ "mulnhh.w %0, %1:b, %2:b" -+ [(set_attr "type" "mulhh") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "machisi3" -+ [(set (match_operand:SI 0 "register_operand" "+r") -+ (plus:SI (mult:SI -+ (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) -+ (match_dup 0)))] -+ "TARGET_DSP" -+ "machh.w %0, %1:b, %2:b" -+ [(set_attr "type" "machh_w") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+ -+ -+(define_insn "mulsidi3" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (mult:DI -+ (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:SI 2 "register_operand" "r"))))] -+ "" -+ "muls.d %0, %1, %2" -+ [(set_attr "type" "mulww_d") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "umulsidi3" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (mult:DI -+ (zero_extend:DI (match_operand:SI 1 "register_operand" "%r")) -+ (zero_extend:DI (match_operand:SI 2 "register_operand" "r"))))] -+ "" -+ "mulu.d %0, %1, %2" -+ [(set_attr "type" "mulww_d") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "*mulaccsi3" -+ [(set (match_operand:SI 0 "register_operand" "+r") -+ (plus:SI (mult:SI (match_operand:SI 1 "register_operand" "%r") -+ (match_operand:SI 2 "register_operand" "r")) -+ (match_dup 0)))] -+ "" -+ "mac %0, %1, %2" -+ [(set_attr "type" "macww_w") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "mulaccsidi3" -+ [(set (match_operand:DI 0 "register_operand" "+r") -+ (plus:DI (mult:DI -+ (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:SI 2 "register_operand" "r"))) -+ (match_dup 0)))] -+ "" -+ "macs.d %0, %1, %2" -+ [(set_attr "type" "macww_d") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+(define_insn "umulaccsidi3" -+ [(set (match_operand:DI 0 "register_operand" "+r") -+ (plus:DI (mult:DI -+ (zero_extend:DI (match_operand:SI 1 "register_operand" "%r")) -+ (zero_extend:DI (match_operand:SI 2 "register_operand" "r"))) -+ (match_dup 0)))] -+ "" -+ "macu.d %0, %1, %2" -+ [(set_attr "type" "macww_d") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+ -+ -+;; Try to avoid Write-After-Write hazards for mul operations -+;; if it can be done -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (mult:SI -+ (sign_extend:SI (match_operand 1 "general_operand" "")) -+ (sign_extend:SI (match_operand 2 "general_operand" "")))) -+ (set (match_dup 0) -+ (match_operator:SI 3 "alu_operator" [(match_dup 0) -+ (match_operand 4 "general_operand" "")]))] -+ "peep2_reg_dead_p(1, operands[2])" -+ [(set (match_dup 5) -+ (mult:SI -+ (sign_extend:SI (match_dup 1)) -+ (sign_extend:SI (match_dup 2)))) -+ (set (match_dup 0) -+ (match_op_dup 3 [(match_dup 5) -+ (match_dup 4)]))] -+ "{operands[5] = gen_rtx_REG(SImode, REGNO(operands[2]));}" -+ ) -+ -+ -+ -+;;============================================================================= -+;; DSP instructions -+;;============================================================================= -+(define_insn "mulsathh_h" -+ [(set (match_operand:HI 0 "register_operand" "=r") -+ (ss_truncate:HI (ashiftrt:SI (mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 15))))] -+ "TARGET_DSP" -+ "mulsathh.h\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulhh")]) -+ -+(define_insn "mulsatrndhh_h" -+ [(set (match_operand:HI 0 "register_operand" "=r") -+ (ss_truncate:HI (ashiftrt:SI -+ (plus:SI (mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 1073741824)) -+ (const_int 15))))] -+ "TARGET_DSP" -+ "mulsatrndhh.h\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulhh")]) -+ -+(define_insn "mulsathh_w" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ss_truncate:SI (ashift:DI (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 1))))] -+ "TARGET_DSP" -+ "mulsathh.w\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulhh")]) -+ -+(define_insn "mulsatwh_w" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ss_truncate:SI (ashiftrt:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 15))))] -+ "TARGET_DSP" -+ "mulsatwh.w\t%0, %1, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+(define_insn "mulsatrndwh_w" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ss_truncate:SI (ashiftrt:DI (plus:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 1073741824)) -+ (const_int 15))))] -+ "TARGET_DSP" -+ "mulsatrndwh.w\t%0, %1, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+(define_insn "macsathh_w" -+ [(set (match_operand:SI 0 "register_operand" "+r") -+ (plus:SI (match_dup 0) -+ (ss_truncate:SI (ashift:DI (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 1)))))] -+ "TARGET_DSP" -+ "macsathh.w\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulhh")]) -+ -+ -+(define_insn "mulwh_d" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (ashift:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 16)))] -+ "TARGET_DSP" -+ "mulwh.d\t%0, %1, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+ -+(define_insn "mulnwh_d" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (ashift:DI (mult:DI (not:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r"))) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 16)))] -+ "TARGET_DSP" -+ "mulnwh.d\t%0, %1, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+(define_insn "macwh_d" -+ [(set (match_operand:DI 0 "register_operand" "+r") -+ (plus:DI (match_dup 0) -+ (ashift:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) -+ (const_int 16))))] -+ "TARGET_DSP" -+ "macwh.d\t%0, %1, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+(define_insn "machh_d" -+ [(set (match_operand:DI 0 "register_operand" "+r") -+ (plus:DI (match_dup 0) -+ (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) -+ (sign_extend:DI (match_operand:HI 2 "register_operand" "r")))))] -+ "TARGET_DSP" -+ "machh.d\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "mulwh")]) -+ -+(define_insn "satadd_w" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ss_plus:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")))] -+ "TARGET_DSP" -+ "satadd.w\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "alu_sat")]) -+ -+(define_insn "satsub_w" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ss_minus:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")))] -+ "TARGET_DSP" -+ "satsub.w\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "alu_sat")]) -+ -+(define_insn "satadd_h" -+ [(set (match_operand:HI 0 "register_operand" "=r") -+ (ss_plus:HI (match_operand:HI 1 "register_operand" "r") -+ (match_operand:HI 2 "register_operand" "r")))] -+ "TARGET_DSP" -+ "satadd.h\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "alu_sat")]) -+ -+(define_insn "satsub_h" -+ [(set (match_operand:HI 0 "register_operand" "=r") -+ (ss_minus:HI (match_operand:HI 1 "register_operand" "r") -+ (match_operand:HI 2 "register_operand" "r")))] -+ "TARGET_DSP" -+ "satsub.h\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none") -+ (set_attr "type" "alu_sat")]) -+ -+ -+;;============================================================================= -+;; smin -+;;----------------------------------------------------------------------------- -+;; Set reg0 to the smallest value of reg1 and reg2. It is used for signed -+;; values in the registers. -+;;============================================================================= -+(define_insn "sminsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (smin:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")))] -+ "" -+ "min %0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+;;============================================================================= -+;; smax -+;;----------------------------------------------------------------------------- -+;; Set reg0 to the largest value of reg1 and reg2. It is used for signed -+;; values in the registers. -+;;============================================================================= -+(define_insn "smaxsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (smax:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")))] -+ "" -+ "max %0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+ -+;;============================================================================= -+;; Logical operations -+;;----------------------------------------------------------------------------- -+ -+;; Split up simple DImode logical operations. Simply perform the logical -+;; operation on the upper and lower halves of the registers. -+(define_split -+ [(set (match_operand:DI 0 "register_operand" "") -+ (match_operator:DI 6 "logical_binary_operator" -+ [(match_operand:DI 1 "register_operand" "") -+ (match_operand:DI 2 "register_operand" "")]))] -+ "reload_completed" -+ [(set (match_dup 0) (match_op_dup:SI 6 [(match_dup 1) (match_dup 2)])) -+ (set (match_dup 3) (match_op_dup:SI 6 [(match_dup 4) (match_dup 5)]))] -+ " -+ { -+ operands[3] = gen_highpart (SImode, operands[0]); -+ operands[0] = gen_lowpart (SImode, operands[0]); -+ operands[4] = gen_highpart (SImode, operands[1]); -+ operands[1] = gen_lowpart (SImode, operands[1]); -+ operands[5] = gen_highpart (SImode, operands[2]); -+ operands[2] = gen_lowpart (SImode, operands[2]); -+ }" -+) -+ -+;;============================================================================= -+;; Logical operations with shifted operand -+;;============================================================================= -+(define_insn "si_lshift" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (logical:SI (match_operator:SI 4 "logical_shift_operator" -+ [(match_operand:SI 2 "register_operand" "r") -+ (match_operand:SI 3 "immediate_operand" "Ku05")]) -+ (match_operand:SI 1 "register_operand" "r")))] -+ "" -+ { -+ if ( GET_CODE(operands[4]) == ASHIFT ) -+ return "\t%0, %1, %2 << %3"; -+ else -+ return "\t%0, %1, %2 >> %3"; -+ } -+ -+ [(set_attr "cc" "set_z")] -+) -+ -+ -+;;************************************************ -+;; Peepholes for detecting logical operantions -+;; with shifted operands -+;;************************************************ -+ -+(define_peephole -+ [(set (match_operand:SI 3 "register_operand" "") -+ (match_operator:SI 5 "logical_shift_operator" -+ [(match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "")])) -+ (set (match_operand:SI 0 "register_operand" "") -+ (logical:SI (match_operand:SI 4 "register_operand" "") -+ (match_dup 3)))] -+ "(dead_or_set_p(insn, operands[3])) || (REGNO(operands[3]) == REGNO(operands[0]))" -+ { -+ if ( GET_CODE(operands[5]) == ASHIFT ) -+ return "\t%0, %4, %1 << %2"; -+ else -+ return "\t%0, %4, %1 >> %2"; -+ } -+ [(set_attr "cc" "set_z")] -+ ) -+ -+(define_peephole -+ [(set (match_operand:SI 3 "register_operand" "") -+ (match_operator:SI 5 "logical_shift_operator" -+ [(match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "")])) -+ (set (match_operand:SI 0 "register_operand" "") -+ (logical:SI (match_dup 3) -+ (match_operand:SI 4 "register_operand" "")))] -+ "(dead_or_set_p(insn, operands[3])) || (REGNO(operands[3]) == REGNO(operands[0]))" -+ { -+ if ( GET_CODE(operands[5]) == ASHIFT ) -+ return "\t%0, %4, %1 << %2"; -+ else -+ return "\t%0, %4, %1 >> %2"; -+ } -+ [(set_attr "cc" "set_z")] -+ ) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (match_operator:SI 5 "logical_shift_operator" -+ [(match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "")])) -+ (set (match_operand:SI 3 "register_operand" "") -+ (logical:SI (match_operand:SI 4 "register_operand" "") -+ (match_dup 0)))] -+ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[3]) == REGNO(operands[0]))" -+ -+ [(set (match_dup 3) -+ (logical:SI (match_op_dup:SI 5 [(match_dup 1) (match_dup 2)]) -+ (match_dup 4)))] -+ -+ "" -+) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (match_operator:SI 5 "logical_shift_operator" -+ [(match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" "")])) -+ (set (match_operand:SI 3 "register_operand" "") -+ (logical:SI (match_dup 0) -+ (match_operand:SI 4 "register_operand" "")))] -+ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[3]) == REGNO(operands[0]))" -+ -+ [(set (match_dup 3) -+ (logical:SI (match_op_dup:SI 5 [(match_dup 1) (match_dup 2)]) -+ (match_dup 4)))] -+ -+ "" -+) -+ -+ -+;;============================================================================= -+;; and -+;;----------------------------------------------------------------------------- -+;; Store the result after a bitwise logical-and between reg0 and reg2 in reg0. -+;;============================================================================= -+ -+(define_insn "andnsi" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (and:SI (match_operand:SI 1 "register_operand" "0") -+ (not:SI (match_operand:SI 2 "register_operand" "r"))))] -+ "" -+ "andn %0, %2" -+ [(set_attr "cc" "set_z") -+ (set_attr "length" "2")] -+) -+ -+ -+ -+ -+(define_insn "andsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r, r, r, r") -+ (and:SI (match_operand:SI 1 "register_operand" "%0, r, 0, r") -+ (match_operand:SI 2 "nonmemory_operand" "r, M, i, r")))] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "and\t%0, %2"; -+ case 1: -+ { -+ int i, first_set = -1; -+ /* Search for first bit set in mask */ -+ for ( i = 31; i >= 0; --i ) -+ if ( INTVAL(operands[2]) & (1 << i) ){ -+ first_set = i; -+ break; -+ } -+ operands[2] = gen_rtx_CONST_INT(SImode, first_set + 1); -+ return "bfextu\t%0, %1, 0, %2"; -+ } -+ case 2: -+ if ( one_bit_cleared_operand(operands[2], VOIDmode) ){ -+ int bitpos; -+ for ( bitpos = 0; bitpos < 32; bitpos++ ) -+ if ( !(INTVAL(operands[2]) & (1 << bitpos)) ) -+ break; -+ operands[2] = gen_rtx_CONST_INT(SImode, bitpos); -+ return "cbr\t%0, %2"; -+ } else if ( (INTVAL(operands[2]) >= 0) && -+ (INTVAL(operands[2]) <= 65535) ) -+ return "andl\t%0, %2, COH"; -+ else if ( (INTVAL(operands[2]) < 0) && -+ (INTVAL(operands[2]) >= -65536 ) ) -+ return "andl\t%0, lo(%2)"; -+ else if ( ((INTVAL(operands[2]) & 0xffff) == 0xffff) ) -+ return "andh\t%0, hi(%2)"; -+ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) -+ return "andh\t%0, hi(%2), COH"; -+ else -+ return "andh\t%0, hi(%2)\;andl\t%0, lo(%2)"; -+ case 3: -+ return "and\t%0, %1, %2"; -+ default: -+ abort(); -+ } -+ } -+ -+ [(set_attr "length" "2,4,8,4") -+ (set_attr "cc" "set_z")]) -+ -+ -+(define_insn "anddi3" -+ [(set (match_operand:DI 0 "register_operand" "=&r,&r") -+ (and:DI (match_operand:DI 1 "register_operand" "%0,r") -+ (match_operand:DI 2 "register_operand" "r,r")))] -+ "" -+ "#" -+ [(set_attr "length" "8") -+ (set_attr "cc" "clobber")] -+) -+ -+;;============================================================================= -+;; or -+;;----------------------------------------------------------------------------- -+;; Store the result after a bitwise inclusive-or between reg0 and reg2 in reg0. -+;;============================================================================= -+ -+(define_insn "iorsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (ior:SI (match_operand:SI 1 "register_operand" "%0,0,r" ) -+ (match_operand:SI 2 "nonmemory_operand" "r ,i,r")))] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "or\t%0, %2"; -+ case 1: -+ if ( one_bit_set_operand(operands[2], VOIDmode) ){ -+ int bitpos; -+ for (bitpos = 0; bitpos < 32; bitpos++) -+ if (INTVAL(operands[2]) & (1 << bitpos)) -+ break; -+ operands[2] = gen_rtx_CONST_INT( SImode, bitpos); -+ return "sbr\t%0, %2"; -+ } else if ( (INTVAL(operands[2]) >= 0) && -+ (INTVAL(operands[2]) <= 65535) ) -+ return "orl\t%0, %2"; -+ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) -+ return "orh\t%0, hi(%2)"; -+ else -+ return "orh\t%0, hi(%2)\;orl\t%0, lo(%2)"; -+ case 2: -+ return "or\t%0, %1, %2"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,8,4") -+ (set_attr "cc" "set_z")]) -+ -+ -+;(define_insn "iorsi3" -+; [(set (match_operand:SI 0 "register_operand" "=r, r, r") -+; (ior:SI (match_operand:SI 1 "avr32_logical_insn_operand" "r, r, rA" ) -+; (match_operand:SI 2 "register_operand" "0, i, r")))] -+; "" -+; { -+; switch (which_alternative){ -+; case 0: -+; return "or %0, %2"; -+; case 1: -+; if ( one_bit_set_operand(operands[2], VOIDmode) ){ -+; int i, bitpos; -+; for ( i = 0; i < 32; i++ ) -+; if ( INTVAL(operands[2]) & (1 << i) ){ -+; bitpos = i; -+; break; -+; } -+; operands[2] = gen_rtx_CONST_INT( SImode, bitpos); -+; return "sbr %0, %2"; -+; } else if ( (INTVAL(operands[2]) >= 0) && -+; (INTVAL(operands[2]) <= 65535) ) -+; return "orl %0, %2"; -+; else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) -+; return "orh %0, hi(%2)"; -+; else -+; return "orh %0, hi(%2)\;orl %0, lo(%2)"; -+; case 2: -+; return "or %0, %2, %1"; -+; } -+; } -+; [(set_attr "length" "2,8,4") -+; (set_attr "cc" "set_z")]) -+ -+(define_insn "iordi3" -+ [(set (match_operand:DI 0 "register_operand" "=&r,&r") -+ (ior:DI (match_operand:DI 1 "register_operand" "%0,r") -+ (match_operand:DI 2 "register_operand" "r,r")))] -+ "" -+ "#" -+ [(set_attr "length" "8") -+ (set_attr "cc" "clobber")] -+) -+ -+;;============================================================================= -+;; xor bytes -+;;----------------------------------------------------------------------------- -+;; Store the result after a bitwise exclusive-or between reg0 and reg2 in reg0. -+;;============================================================================= -+ -+(define_insn "xorsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (xor:SI (match_operand:SI 1 "register_operand" "0,0,r") -+ (match_operand:SI 2 "nonmemory_operand" "r,i,r")))] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "eor %0, %2"; -+ case 1: -+ if ( (INTVAL(operands[2]) >= 0) && -+ (INTVAL(operands[2]) <= 65535) ) -+ return "eorl %0, %2"; -+ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) -+ return "eorh %0, hi(%2)"; -+ else -+ return "eorh %0, hi(%2)\;eorl %0, lo(%2)"; -+ case 2: -+ return "eor %0, %1, %2"; -+ default: -+ abort(); -+ } -+ } -+ -+ [(set_attr "length" "2,8,4") -+ (set_attr "cc" "set_z")]) -+ -+(define_insn "xordi3" -+ [(set (match_operand:DI 0 "register_operand" "=&r,&r") -+ (xor:DI (match_operand:DI 1 "register_operand" "%0,r") -+ (match_operand:DI 2 "register_operand" "r,r")))] -+ "" -+ "#" -+ [(set_attr "length" "8") -+ (set_attr "cc" "clobber")] -+) -+ -+;;============================================================================= -+;; divmod -+;;----------------------------------------------------------------------------- -+;; Signed division that produces both a quotient and a remainder. -+;;============================================================================= -+(define_expand "divmodsi4" -+ [(parallel [ -+ (parallel [ -+ (set (match_operand:SI 0 "register_operand" "=r") -+ (div:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r"))) -+ (set (match_operand:SI 3 "register_operand" "=r") -+ (mod:SI (match_dup 1) -+ (match_dup 2)))]) -+ (use (match_dup 4))])] -+ "" -+ { -+ if (! no_new_pseudos) { -+ operands[4] = gen_reg_rtx (DImode); -+ -+ emit_insn(gen_divmodsi4_internal(operands[4],operands[1],operands[2])); -+ emit_move_insn(operands[0], gen_rtx_SUBREG( SImode, operands[4], 4)); -+ emit_move_insn(operands[3], gen_rtx_SUBREG( SImode, operands[4], 0)); -+ -+ DONE; -+ } else { -+ FAIL; -+ } -+ -+ }) -+ -+ -+(define_insn "divmodsi4_internal" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (unspec:DI [(match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")] -+ UNSPEC_DIVMODSI4_INTERNAL))] -+ "" -+ "divs %0, %1, %2" -+ [(set_attr "type" "div") -+ (set_attr "cc" "none")]) -+ -+ -+;;============================================================================= -+;; udivmod -+;;----------------------------------------------------------------------------- -+;; Unsigned division that produces both a quotient and a remainder. -+;;============================================================================= -+(define_expand "udivmodsi4" -+ [(parallel [ -+ (parallel [ -+ (set (match_operand:SI 0 "register_operand" "=r") -+ (udiv:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r"))) -+ (set (match_operand:SI 3 "register_operand" "=r") -+ (umod:SI (match_dup 1) -+ (match_dup 2)))]) -+ (use (match_dup 4))])] -+ "" -+ { -+ if (! no_new_pseudos) { -+ operands[4] = gen_reg_rtx (DImode); -+ -+ emit_insn(gen_udivmodsi4_internal(operands[4],operands[1],operands[2])); -+ emit_move_insn(operands[0], gen_rtx_SUBREG( SImode, operands[4], 4)); -+ emit_move_insn(operands[3], gen_rtx_SUBREG( SImode, operands[4], 0)); -+ -+ DONE; -+ } else { -+ FAIL; -+ } -+ }) -+ -+(define_insn "udivmodsi4_internal" -+ [(set (match_operand:DI 0 "register_operand" "=r") -+ (unspec:DI [(match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "register_operand" "r")] -+ UNSPEC_UDIVMODSI4_INTERNAL))] -+ "" -+ "divu %0, %1, %2" -+ [(set_attr "type" "div") -+ (set_attr "cc" "none")]) -+ -+ -+;;============================================================================= -+;; Arithmetic-shift left -+;;----------------------------------------------------------------------------- -+;; Arithmetic-shift reg0 left by reg2 or immediate value. -+;;============================================================================= -+ -+(define_insn "ashlsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (ashift:SI (match_operand:SI 1 "register_operand" "r,0,r") -+ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] -+ "" -+ "@ -+ lsl %0, %1, %2 -+ lsl %0, %2 -+ lsl %0, %1, %2" -+ [(set_attr "length" "4,2,4") -+ (set_attr "cc" "set_ncz")]) -+ -+;;============================================================================= -+;; Arithmetic-shift right -+;;----------------------------------------------------------------------------- -+;; Arithmetic-shift reg0 right by an immediate value. -+;;============================================================================= -+ -+(define_insn "ashrsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (ashiftrt:SI (match_operand:SI 1 "register_operand" "r,0,r") -+ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] -+ "" -+ "@ -+ asr %0, %1, %2 -+ asr %0, %2 -+ asr %0, %1, %2" -+ [(set_attr "length" "4,2,4") -+ (set_attr "cc" "set_ncz")]) -+ -+;;============================================================================= -+;; Logical shift right -+;;----------------------------------------------------------------------------- -+;; Logical shift reg0 right by an immediate value. -+;;============================================================================= -+ -+(define_insn "lshrsi3" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r") -+ (lshiftrt:SI (match_operand:SI 1 "register_operand" "r,0,r") -+ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] -+ "" -+ "@ -+ lsr %0, %1, %2 -+ lsr %0, %2 -+ lsr %0, %1, %2" -+ [(set_attr "length" "4,2,4") -+ (set_attr "cc" "set_ncz")]) -+ -+ -+;;============================================================================= -+;; neg -+;;----------------------------------------------------------------------------- -+;; Negate operand 1 and store the result in operand 0. -+;;============================================================================= -+(define_insn "negsi2" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (neg:SI (match_operand:SI 1 "register_operand" "0")))] -+ "" -+ "neg %0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "set_vncz")]) -+ -+;;============================================================================= -+;; abs -+;;----------------------------------------------------------------------------- -+;; Store the absolute value of operand 1 into operand 0. -+;;============================================================================= -+(define_insn "abssi2" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (abs:SI (match_operand:SI 1 "register_operand" "0")))] -+ "" -+ "abs %0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "set_z")]) -+ -+ -+;;============================================================================= -+;; one_cmpl -+;;----------------------------------------------------------------------------- -+;; Store the bitwise-complement of operand 1 into operand 0. -+;;============================================================================= -+ -+(define_insn "one_cmplsi2" -+ [(set (match_operand:SI 0 "register_operand" "=r,r") -+ (not:SI (match_operand:SI 1 "register_operand" "r,0")))] -+ "" -+ "@ -+ rsub %0, %1, -1 -+ com %0" -+ [(set_attr "length" "4,2") -+ (set_attr "cc" "set_z")]) -+ -+ -+;;============================================================================= -+;; Bit load -+;;----------------------------------------------------------------------------- -+;; Load a bit into Z and C flags -+;;============================================================================= -+(define_insn "bldsi" -+ [(set (cc0) -+ (and:SI (match_operand:SI 0 "register_operand" "r") -+ (match_operand:SI 1 "one_bit_set_operand" "i")))] -+ "" -+ "bld\t%0, %p1" -+ [(set_attr "length" "4") -+ (set_attr "cc" "bld")] -+ ) -+ -+ -+;;============================================================================= -+;; Compare -+;;----------------------------------------------------------------------------- -+;; Compare reg0 with reg1 or an immediate value. -+;;============================================================================= -+ -+(define_expand "cmpqi" -+ [(set (cc0) -+ (compare:QI -+ (match_operand:QI 0 "general_operand" "") -+ (match_operand:QI 1 "general_operand" "")))] -+ "" -+ "{ -+ -+ if ( GET_CODE(operands[0]) != REG -+ && GET_CODE(operands[0]) != SUBREG) -+ operands[0] = force_reg(QImode, operands[0]); -+ -+ -+ if ( GET_CODE(operands[1]) != REG -+ && GET_CODE(operands[1]) != SUBREG ) -+ operands[1] = force_reg(QImode, operands[1]); -+ -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ emit_insn(gen_cmpqi_internal(operands[0], operands[1])); -+ DONE; -+ }" -+) -+ -+ -+(define_insn "cmpqi_internal" -+ [(set (cc0) -+ (compare:QI -+ (match_operand:QI 0 "register_operand" "r") -+ (match_operand:QI 1 "register_operand" "r")))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), QImode, operands[0], operands[1])); -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "cc" "compare")]) -+ -+(define_expand "cmphi" -+ [(set (cc0) -+ (compare:HI -+ (match_operand:HI 0 "register_operand" "") -+ (match_operand:HI 1 "register_operand" "")))] -+ "" -+ "{ -+ -+ //if ( (GET_CODE(operands[0]) == REG -+ // || GET_CODE(operands[0]) == SUBREG) -+ // && (GET_CODE(operands[1]) == CONST_INT -+ // && avr32_const_ok_for_constraint_p (INTVAL(operands[1]), 'K', \"Ks21\")) ){ -+ // operands[0] = convert_to_mode(SImode, operands[0], 0); -+ // avr32_compare_op0 = operands[0]; -+ // avr32_compare_op1 = operands[1]; -+ // emit_insn(gen_cmpsi_internal(operands[0], operands[1])); -+ // DONE; -+ //} -+ -+ if ( GET_CODE(operands[0]) != REG -+ && GET_CODE(operands[0]) != SUBREG ) -+ operands[0] = force_reg(HImode, operands[0]); -+ -+ -+ if ( GET_CODE(operands[1]) != REG -+ && GET_CODE(operands[1]) != SUBREG) -+ operands[1] = force_reg(HImode, operands[1]); -+ -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ }" -+) -+ -+ -+(define_insn "cmphi_internal" -+ [(set (cc0) -+ (compare:HI -+ (match_operand:HI 0 "register_operand" "r") -+ (match_operand:HI 1 "register_operand" "r")))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), HImode, operands[0], operands[1])); -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "cc" "compare")]) -+ -+ -+(define_expand "cmpsi" -+ [(set (cc0) -+ (compare:SI -+ (match_operand:SI 0 "register_operand" "") -+ (match_operand:SI 1 "register_immediate_operand" "")))] -+ "" -+ "{ -+ if ( GET_CODE(operands[0]) != REG -+ && GET_CODE(operands[0]) != SUBREG ) -+ operands[0] = force_reg(SImode, operands[0]); -+ -+ if ( GET_CODE(operands[1]) != REG -+ && GET_CODE(operands[1]) != SUBREG -+ && GET_CODE(operands[1]) != CONST_INT ) -+ operands[1] = force_reg(SImode, operands[1]); -+ -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ -+ }" -+) -+ -+ -+(define_insn "cmpsi_internal" -+ [(set (cc0) -+ (compare:SI -+ (match_operand:SI 0 "register_operand" "r, r, r") -+ (match_operand:SI 1 "register_immediate_operand" "r, Ks06, Ks21")))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), SImode, operands[0], operands[1])); -+ return ""; -+ } -+ -+ [(set_attr "length" "2,2,4") -+ (set_attr "cc" "compare")]) -+ -+ -+(define_expand "cmpdi" -+ [(set (cc0) -+ (compare:DI -+ (match_operand:DI 0 "register_operand" "") -+ (match_operand:DI 1 "register_immediate_operand" "")))] -+ "" -+ { -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ } -+) -+ -+(define_insn "cmpdi_internal" -+ [(set (cc0) -+ (compare:DI -+ (match_operand:DI 0 "register_operand" "r") -+ (match_operand:DI 1 "register_immediate_operand" "rKu20")))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), DImode, operands[0], operands[1])); -+ return ""; -+ } -+ -+ [(set_attr "length" "6") -+ (set_attr "type" "alu2") -+ (set_attr "cc" "compare")]) -+ -+ -+ -+;;============================================================================= -+;; Test if zero -+;;----------------------------------------------------------------------------- -+;; Compare reg against zero and set the condition codes. -+;;============================================================================= -+ -+ -+(define_expand "tstsi" -+ [(set (cc0) -+ (match_operand:SI 0 "register_operand" ""))] -+ "" -+ { -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = const0_rtx; -+ } -+) -+ -+(define_insn "tstsi_internal" -+ [(set (cc0) -+ (match_operand:SI 0 "register_operand" "r"))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), SImode, operands[0], const0_rtx)); -+ -+ return ""; -+ } -+ [(set_attr "length" "2") -+ (set_attr "cc" "compare")]) -+ -+ -+(define_expand "tstdi" -+ [(set (cc0) -+ (match_operand:DI 0 "register_operand" ""))] -+ "" -+ { -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = const0_rtx; -+ } -+) -+ -+(define_insn "tstdi_internal" -+ [(set (cc0) -+ (match_operand:DI 0 "register_operand" "r"))] -+ "" -+ { -+ set_next_insn_cond(insn, -+ avr32_output_cmp(get_next_insn_cond(insn), DImode, operands[0], const0_rtx)); -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "type" "alu2") -+ (set_attr "cc" "compare")]) -+ -+ -+ -+;;============================================================================= -+;; Convert operands -+;;----------------------------------------------------------------------------- -+;; -+;;============================================================================= -+(define_insn "truncdisi2" -+ [(set (match_operand:SI 0 "general_operand" "") -+ (truncate:SI (match_operand:DI 1 "general_operand" "")))] -+ "" -+ "truncdisi2") -+ -+;;============================================================================= -+;; Extend -+;;----------------------------------------------------------------------------- -+;; -+;;============================================================================= -+ -+ -+(define_insn "extendhisi2" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") -+ (sign_extend:SI (match_operand:HI 1 "nonimmediate_operand" "0,r,,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "casts.h\t%0"; -+ case 1: -+ return "bfexts\t%0, %1, 0, 16"; -+ case 2: -+ case 3: -+ return "ld.sh\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz,set_ncz,none,none") -+ (set_attr "type" "alu,alu,load_rm,load_rm")]) -+ -+(define_insn "extendqisi2" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") -+ (sign_extend:SI (match_operand:QI 1 "extendqi_operand" "0,r,RKu00,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "casts.b\t%0"; -+ case 1: -+ return "bfexts\t%0, %1, 0, 8"; -+ case 2: -+ case 3: -+ return "ld.sb\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz,set_ncz,none,none") -+ (set_attr "type" "alu,alu,load_rm,load_rm")]) -+ -+(define_insn "extendqihi2" -+ [(set (match_operand:HI 0 "register_operand" "=r,r,r,r") -+ (sign_extend:HI (match_operand:QI 1 "extendqi_operand" "0,r,RKu00,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "casts.b\t%0"; -+ case 1: -+ return "bfexts\t%0, %1, 0, 8"; -+ case 2: -+ case 3: -+ return "ld.sb\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz,set_ncz,none,none") -+ (set_attr "type" "alu,alu,load_rm,load_rm")]) -+ -+ -+;;============================================================================= -+;; Zero-extend -+;;----------------------------------------------------------------------------- -+;; -+;;============================================================================= -+ -+(define_insn "zero_extendhisi2" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") -+ (zero_extend:SI (match_operand:HI 1 "nonimmediate_operand" "0,r,,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "castu.h\t%0"; -+ case 1: -+ return "bfextu\t%0, %1, 0, 16"; -+ case 2: -+ case 3: -+ return "ld.uh\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz,set_ncz,none,none") -+ (set_attr "type" "alu,alu,load_rm,load_rm")]) -+ -+(define_insn "zero_extendqisi2" -+ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") -+ (zero_extend:SI (match_operand:QI 1 "nonimmediate_operand" "0,r,,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "castu.b\t%0"; -+ case 1: -+ return "bfextu\t%0, %1, 0, 8"; -+ case 2: -+ case 3: -+ return "ld.ub\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz, set_ncz, none, none") -+ (set_attr "type" "alu, alu, load_rm, load_rm")]) -+ -+(define_insn "zero_extendqihi2" -+ [(set (match_operand:HI 0 "register_operand" "=r,r,r,r") -+ (zero_extend:HI (match_operand:QI 1 "nonimmediate_operand" "0,r,,m")))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ return "castu.b\t%0"; -+ case 1: -+ return "bfextu\t%0, %1, 0, 8"; -+ case 2: -+ case 3: -+ return "ld.ub\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,2,4") -+ (set_attr "cc" "set_ncz, set_ncz, none, none") -+ (set_attr "type" "alu, alu, load_rm, load_rm")]) -+ -+ -+ -+;;============================================================================= -+;; Conditional set register -+;; sr{cond4} rd -+;;----------------------------------------------------------------------------- -+ -+;;Because of the same issue as with conditional moves and adds we must -+;;not separate the compare instrcution from the scc instruction as -+;;they might be sheduled "badly". -+ -+(define_insn "s" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (any_cond:SI (cc0) -+ (const_int 0)))] -+ "" -+ "sr\t%0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "none")]) -+ -+(define_insn "smi" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (unspec:SI [(cc0) -+ (const_int 0)] UNSPEC_COND_MI))] -+ "" -+ "srmi\t%0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "none")]) -+ -+(define_insn "spl" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (unspec:SI [(cc0) -+ (const_int 0)] UNSPEC_COND_PL))] -+ "" -+ "srpl\t%0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "none")]) -+ -+ -+;;============================================================================= -+;; Conditional branch -+;;----------------------------------------------------------------------------- -+;; Branch to label if the specified condition codes are set. -+;;============================================================================= -+; branch if negative -+(define_insn "bmi" -+ [(set (pc) -+ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_MI) -+ (label_ref (match_operand 0 "" "")) -+ (pc)))] -+ "" -+ "brmi %0" -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+(define_insn "*bmi-reverse" -+ [(set (pc) -+ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_MI) -+ (pc) -+ (label_ref (match_operand 0 "" ""))))] -+ "" -+ "brpl %0" -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+; branch if positive -+(define_insn "bpl" -+ [(set (pc) -+ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_PL) -+ (label_ref (match_operand 0 "" "")) -+ (pc)))] -+ "" -+ "brpl %0" -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+(define_insn "*bpl-reverse" -+ [(set (pc) -+ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_PL) -+ (pc) -+ (label_ref (match_operand 0 "" ""))))] -+ "" -+ "brmi %0" -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+; branch if equal -+(define_insn "b" -+ [(set (pc) -+ (if_then_else (any_cond:CC (cc0) -+ (const_int 0)) -+ (label_ref (match_operand 0 "" "")) -+ (pc)))] -+ "" -+ "br %0 " -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+ -+(define_insn "*b-reverse" -+ [(set (pc) -+ (if_then_else (any_cond:CC (cc0) -+ (const_int 0)) -+ (pc) -+ (label_ref (match_operand 0 "" ""))))] -+ "" -+ "br %0 " -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) -+ (le (minus (pc) (match_dup 0)) (const_int 256))) -+ (const_int 2)] ; use compact branch -+ (const_int 4))) ; use extended branch -+ (set_attr "cc" "none")]) -+ -+ -+ -+;============================================================================= -+; Conditional Add/Subtract -+;----------------------------------------------------------------------------- -+; sub{cond4} Rd, imm -+;============================================================================= -+ -+ -+(define_expand "addcc" -+ [(set (match_operand:ADDCC 0 "register_operand" "") -+ (if_then_else:ADDCC (match_operator 1 "avr32_comparison_operator" -+ [(match_dup 4) -+ (match_dup 5)]) -+ (match_operand:ADDCC 2 "register_operand" "") -+ (plus:ADDCC -+ (match_dup 2) -+ (match_operand:ADDCC 3 "avr32_cond_immediate_operand" ""))))] -+ "" -+ { -+ /* Delete compare instruction as it is merged into this instruction */ -+ remove_insn (get_last_insn_anywhere ()); -+ -+ operands[4] = avr32_compare_op0; -+ operands[5] = avr32_compare_op1; -+ } -+ ) -+ -+ -+(define_insn "addcc_cmp" -+ [(set (match_operand:ADDCC 0 "register_operand" "=r") -+ (if_then_else:ADDCC (match_operator 1 "avr32_comparison_operator" -+ [(match_operand:CMP 4 "register_operand" "r") -+ (match_operand:CMP 5 "" "")]) -+ (match_operand:ADDCC 2 "register_operand" "0") -+ (plus:ADDCC -+ (match_dup 2) -+ (match_operand:ADDCC 3 "avr32_cond_immediate_operand" "Is08"))))] -+ "" -+ { -+ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); -+ -+ return "sub%i1\t%0, -%3"; -+ } -+ [(set_attr "length" "8") -+ (set_attr "cc" "cmp_cond_insn")]) -+ -+;============================================================================= -+; Conditional Move -+;----------------------------------------------------------------------------- -+; mov{cond4} Rd, (Rs/imm) -+;============================================================================= -+(define_expand "movcc" -+ [(set (match_operand:MOVCC 0 "register_operand" "") -+ (if_then_else:MOVCC (match_operator 1 "avr32_comparison_operator" -+ [(match_dup 4) -+ (match_dup 5)]) -+ (match_operand:MOVCC 2 "avr32_cond_register_immediate_operand" "") -+ (match_operand:MOVCC 3 "avr32_cond_register_immediate_operand" "")))] -+ "" -+ { -+ /* Delete compare instruction as it is merged into this instruction */ -+ remove_insn (get_last_insn_anywhere ()); -+ -+ operands[4] = avr32_compare_op0; -+ operands[5] = avr32_compare_op1; -+ } -+ ) -+ -+ -+(define_insn "movcc_cmp" -+ [(set (match_operand:MOVCC 0 "register_operand" "=r,r,r") -+ (if_then_else:MOVCC (match_operator 1 "avr32_comparison_operator" -+ [(match_operand:CMP 4 "register_operand" "r,r,r") -+ (match_operand:CMP 5 "" ",,")]) -+ (match_operand:MOVCC 2 "avr32_cond_register_immediate_operand" "0, rKs08,rKs08") -+ (match_operand:MOVCC 3 "avr32_cond_register_immediate_operand" "rKs08,0,rKs08")))] -+ "" -+ { -+ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); -+ -+ switch( which_alternative ){ -+ case 0: -+ return "mov%i1 %0, %3"; -+ case 1: -+ return "mov%1 %0, %2"; -+ case 2: -+ return "mov%1 %0, %2\;mov%i1 %0, %3"; -+ default: -+ abort(); -+ } -+ -+ -+ } -+ [(set_attr "length" "8,8,12") -+ (set_attr "cc" "cmp_cond_insn")]) -+ -+ -+;;============================================================================= -+;; jump -+;;----------------------------------------------------------------------------- -+;; Jump inside a function; an unconditional branch to a label. -+;;============================================================================= -+(define_insn "jump" -+ [(set (pc) -+ (label_ref (match_operand 0 "" "")))] -+ "" -+ { -+ if (get_attr_length(insn) > 4) -+ return "Can't jump this far"; -+ return (get_attr_length(insn) == 2 ? -+ "rjmp %0" : "bral %0"); -+ } -+ [(set_attr "type" "branch") -+ (set (attr "length") -+ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 1022)) -+ (le (minus (pc) (match_dup 0)) (const_int 1024))) -+ (const_int 2) ; use rjmp -+ (le (match_dup 0) (const_int 1048575)) -+ (const_int 4)] ; use bral -+ (const_int 8))) ; do something else -+ (set_attr "cc" "none")]) -+ -+;;============================================================================= -+;; call -+;;----------------------------------------------------------------------------- -+;; Subroutine call instruction returning no value. -+;;============================================================================= -+(define_insn "call_internal" -+ [(parallel [(call (mem:SI (match_operand:SI 0 "avr32_call_operand" "r,U,T,W")) -+ (match_operand 1 "" "")) -+ (clobber (reg:SI LR_REGNUM))])] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "icall\t%0"; -+ case 1: -+ return "rcall\t%0"; -+ case 2: -+ return "mcall\t%0"; -+ case 3: -+ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) -+ return "call\t%0"; -+ else -+ return "mcall\tr6[%0@got]"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "type" "call") -+ (set_attr "length" "2,4,4,10") -+ (set_attr "cc" "clobber")]) -+ -+ -+(define_expand "call" -+ [(parallel [(call (match_operand:SI 0 "" "") -+ (match_operand 1 "" "")) -+ (clobber (reg:SI LR_REGNUM))])] -+ "" -+ { -+ rtx call_address; -+ if ( GET_CODE(operands[0]) != MEM ) -+ FAIL; -+ -+ call_address = XEXP(operands[0], 0); -+ -+ /* If assembler supports call pseudo insn and the call -+ address is a symbol then nothing special needs to be done. */ -+ if ( TARGET_HAS_ASM_ADDR_PSEUDOS -+ && (GET_CODE(call_address) == SYMBOL_REF) ){ -+ /* We must however mark the function as using the GOT if -+ flag_pic is set, since the call insn might turn into -+ a mcall using the GOT ptr register. */ -+ if ( flag_pic ){ -+ current_function_uses_pic_offset_table = 1; -+ emit_call_insn(gen_call_internal(call_address, operands[1])); -+ DONE; -+ } -+ } else { -+ if ( flag_pic && -+ GET_CODE(call_address) == SYMBOL_REF ){ -+ current_function_uses_pic_offset_table = 1; -+ emit_call_insn(gen_call_internal(call_address, operands[1])); -+ DONE; -+ } -+ -+ if ( !SYMBOL_REF_RCALL_FUNCTION_P(operands[0]) ){ -+ if ( optimize_size && -+ GET_CODE(call_address) == SYMBOL_REF ){ -+ call_address = force_const_mem(SImode, call_address); -+ } else { -+ call_address = force_reg(SImode, call_address); -+ } -+ } -+ } -+ emit_call_insn(gen_call_internal(call_address, operands[1])); -+ DONE; -+ } -+) -+ -+;;============================================================================= -+;; call_value -+;;----------------------------------------------------------------------------- -+;; Subrutine call instruction returning a value. -+;;============================================================================= -+(define_expand "call_value" -+ [(parallel [(set (match_operand:SI 0 "" "") -+ (call (match_operand:SI 1 "" "") -+ (match_operand 2 "" ""))) -+ (clobber (reg:SI LR_REGNUM))])] -+ "" -+ { -+ rtx call_address; -+ if ( GET_CODE(operands[1]) != MEM ) -+ FAIL; -+ -+ call_address = XEXP(operands[1], 0); -+ -+ /* If assembler supports call pseudo insn and the call -+ address is a symbol then nothing special needs to be done. */ -+ if ( TARGET_HAS_ASM_ADDR_PSEUDOS -+ && (GET_CODE(call_address) == SYMBOL_REF) ){ -+ /* We must however mark the function as using the GOT if -+ flag_pic is set, since the call insn might turn into -+ a mcall using the GOT ptr register. */ -+ if ( flag_pic ) { -+ current_function_uses_pic_offset_table = 1; -+ emit_call_insn(gen_call_value_internal(operands[0], call_address, operands[2])); -+ DONE; -+ } -+ } else { -+ if ( flag_pic && -+ GET_CODE(call_address) == SYMBOL_REF ){ -+ current_function_uses_pic_offset_table = 1; -+ emit_call_insn(gen_call_value_internal(operands[0], call_address, operands[2])); -+ DONE; -+ } -+ -+ if ( !SYMBOL_REF_RCALL_FUNCTION_P(operands[1]) ){ -+ if ( optimize_size && -+ GET_CODE(call_address) == SYMBOL_REF){ -+ call_address = force_const_mem(SImode, call_address); -+ } else { -+ call_address = force_reg(SImode, call_address); -+ } -+ } -+ } -+ emit_call_insn(gen_call_value_internal(operands[0], call_address, -+ operands[2])); -+ DONE; -+ -+ }) -+ -+(define_insn "call_value_internal" -+ [(parallel [(set (match_operand 0 "register_operand" "=r,r,r,r") -+ (call (mem:SI (match_operand:SI 1 "avr32_call_operand" "r,U,T,W")) -+ (match_operand 2 "" ""))) -+ (clobber (reg:SI LR_REGNUM))])] -+ ;; Operand 2 not used on the AVR32. -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "icall\t%1"; -+ case 1: -+ return "rcall\t%1"; -+ case 2: -+ return "mcall\t%1"; -+ case 3: -+ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) -+ return "call\t%1"; -+ else -+ return "mcall\tr6[%1@got]"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "type" "call") -+ (set_attr "length" "2,4,4,10") -+ (set_attr "cc" "call_set")]) -+ -+ -+;;============================================================================= -+;; untyped_call -+;;----------------------------------------------------------------------------- -+;; Subrutine call instruction returning a value of any type. -+;; The code is copied from m68k.md (except gen_blockage is removed) -+;; Fixme! -+;;============================================================================= -+(define_expand "untyped_call" -+ [(parallel [(call (match_operand 0 "avr32_call_operand" "") -+ (const_int 0)) -+ (match_operand 1 "" "") -+ (match_operand 2 "" "")])] -+ "" -+ { -+ int i; -+ -+ emit_call_insn (GEN_CALL (operands[0], const0_rtx, NULL, const0_rtx)); -+ -+ for (i = 0; i < XVECLEN (operands[2], 0); i++) { -+ rtx set = XVECEXP (operands[2], 0, i); -+ emit_move_insn (SET_DEST (set), SET_SRC (set)); -+ } -+ -+ /* The optimizer does not know that the call sets the function value -+ registers we stored in the result block. We avoid problems by -+ claiming that all hard registers are used and clobbered at this -+ point. */ -+ emit_insn (gen_blockage ()); -+ -+ DONE; -+ }) -+ -+ -+;;============================================================================= -+;; return -+;;============================================================================= -+ -+(define_insn "return" -+ [(return)] -+ "USE_RETURN_INSN (FALSE)" -+ { -+ avr32_output_return_instruction(TRUE, FALSE, NULL, NULL); -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "type" "call")] -+ ) -+ -+(define_insn "return_cond" -+ [(set (pc) -+ (if_then_else (match_operand 0 "avr32_comparison_operand" "") -+ (return) -+ (pc)))] -+ "USE_RETURN_INSN (TRUE)" -+ "ret%0\tr12"; -+ [(set_attr "type" "call")]) -+ -+ -+(define_insn "return_imm" -+ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) -+ (use (reg RETVAL_REGNUM)) -+ (return)])] -+ "USE_RETURN_INSN (FALSE) && -+ ((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" -+ { -+ avr32_output_return_instruction(TRUE, FALSE, NULL, operands[0]); -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "type" "call")] -+ ) -+ -+(define_insn "return_imm_cond" -+ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) -+ (use (reg RETVAL_REGNUM)) -+ (set (pc) -+ (if_then_else (match_operand 1 "avr32_comparison_operand" "") -+ (return) -+ (pc)))])] -+ "USE_RETURN_INSN (TRUE) && -+ ((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" -+ "ret%1\t%0"; -+ [(set_attr "type" "call")] -+ ) -+ -+(define_insn "return_reg" -+ [(set (reg RETVAL_REGNUM) (match_operand:MOVM 0 "register_operand" "r")) -+ (use (reg RETVAL_REGNUM)) -+ (return)] -+ "USE_RETURN_INSN (TRUE)" -+ "retal %0"; -+ [(set_attr "type" "call")] -+ ) -+ -+(define_insn "return_reg_cond" -+ [(set (reg RETVAL_REGNUM) (match_operand:MOVM 0 "register_operand" "r")) -+ (use (reg RETVAL_REGNUM)) -+ (set (pc) -+ (if_then_else (match_operator 1 "avr32_comparison_operator" -+ [(cc0) (const_int 0)]) -+ (return) -+ (pc)))] -+ "USE_RETURN_INSN (TRUE)" -+ "ret%1\t%0"; -+ [(set_attr "type" "call")]) -+ -+;;============================================================================= -+;; nop -+;;----------------------------------------------------------------------------- -+;; No-op instruction. -+;;============================================================================= -+(define_insn "nop" -+ [(const_int 0)] -+ "" -+ "nop" -+ [(set_attr "length" "2") -+ (set_attr "type" "alu") -+ (set_attr "cc" "none")]) -+ -+;;============================================================================= -+;; nonlocal_goto_receiver -+;;----------------------------------------------------------------------------- -+;; For targets with a return stack we must make sure to flush the return stack -+;; since it will be corrupt after a nonlocal goto. -+;;============================================================================= -+(define_expand "nonlocal_goto_receiver" -+ [(const_int 0)] -+ "TARGET_RETURN_STACK" -+ " -+ { -+ emit_insn ( gen_frs() ); -+ DONE; -+ } -+ " -+ ) -+ -+ -+;;============================================================================= -+;; builtin_setjmp_receiver -+;;----------------------------------------------------------------------------- -+;; For pic code we need to reload the pic register. -+;; For targets with a return stack we must make sure to flush the return stack -+;; since it will probably be corrupted. -+;;============================================================================= -+(define_expand "builtin_setjmp_receiver" -+ [(label_ref (match_operand 0 "" ""))] -+ "flag_pic" -+ " -+ { -+ if ( TARGET_RETURN_STACK ) -+ emit_insn ( gen_frs() ); -+ avr32_load_pic_register (); -+ DONE; -+ } -+ " -+) -+ -+ -+;;============================================================================= -+;; indirect_jump -+;;----------------------------------------------------------------------------- -+;; Jump to an address in reg or memory. -+;;============================================================================= -+(define_expand "indirect_jump" -+ [(set (pc) -+ (match_operand:SI 0 "general_operand" ""))] -+ "" -+ { -+ /* One of the ops has to be in a register. */ -+ if ( (flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS ) -+ && !avr32_legitimate_pic_operand_p(operands[0]) ) -+ operands[0] = legitimize_pic_address (operands[0], SImode, 0); -+ else if ( flag_pic && avr32_address_operand(operands[0], GET_MODE(operands[0])) ) -+ /* If we have an address operand then this function uses the pic register. */ -+ current_function_uses_pic_offset_table = 1; -+ }) -+ -+ -+(define_insn "indirect_jump_internal" -+ [(set (pc) -+ (match_operand:SI 0 "general_operand" "r,m,W"))] -+ "" -+ { -+ switch( which_alternative ){ -+ case 0: -+ return "mov\tpc, %0"; -+ case 1: -+ if ( avr32_const_pool_ref_operand(operands[0], GET_MODE(operands[0])) ) -+ return "lddpc\tpc, %0"; -+ else -+ return "ld.w\tpc, %0"; -+ case 2: -+ if ( flag_pic ) -+ return "ld.w\tpc, r6[%0@got]"; -+ else -+ return "lda.w\tpc, %0"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "2,4,8") -+ (set_attr "type" "call,call,call") -+ (set_attr "cc" "none,none,clobber")]) -+ -+ -+;;============================================================================= -+;; casesi and tablejump -+;;============================================================================= -+(define_insn "tablejump_add" -+ [(set (pc) -+ (plus:SI (match_operand:SI 0 "register_operand" "r") -+ (mult:SI (match_operand:SI 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku04" )))) -+ (use (label_ref (match_operand 3 "" "")))] -+ "flag_pic && -+ ((INTVAL(operands[2]) == 0) || (INTVAL(operands[2]) == 2) || -+ (INTVAL(operands[2]) == 4) || (INTVAL(operands[2]) == 8))" -+ "add\tpc, %0, %1 << %p2" -+ [(set_attr "length" "4") -+ (set_attr "cc" "clobber")]) -+ -+(define_insn "tablejump_insn" -+ [(set (pc) (match_operand:SI 0 "memory_operand" "m")) -+ (use (label_ref (match_operand 1 "" "")))] -+ "!flag_pic" -+ "ld.w\tpc, %0" -+ [(set_attr "length" "4") -+ (set_attr "type" "call") -+ (set_attr "cc" "none")]) -+ -+(define_expand "casesi" -+ [(match_operand:SI 0 "register_operand" "") ; index to jump on -+ (match_operand:SI 1 "const_int_operand" "") ; lower bound -+ (match_operand:SI 2 "const_int_operand" "") ; total range -+ (match_operand:SI 3 "" "") ; table label -+ (match_operand:SI 4 "" "")] ; Out of range label -+ "" -+ " -+ { -+ rtx reg; -+ rtx index = operands[0]; -+ rtx low_bound = operands[1]; -+ rtx range = operands[2]; -+ rtx table_label = operands[3]; -+ rtx oor_label = operands[4]; -+ -+ if (low_bound != const0_rtx) -+ { -+ if (!avr32_const_ok_for_constraint_p(INTVAL (low_bound), 'I', \"Is21\")){ -+ reg = force_reg(SImode, GEN_INT (INTVAL (low_bound))); -+ emit_insn (gen_subsi3 (reg, index, -+ reg)); -+ } else { -+ reg = gen_reg_rtx (SImode); -+ emit_insn (gen_addsi3 (reg, index, -+ GEN_INT (-INTVAL (low_bound)))); -+ } -+ index = reg; -+ } -+ -+ if (!avr32_const_ok_for_constraint_p (INTVAL (range), 'K', \"Ks21\")) -+ range = force_reg (SImode, range); -+ -+ emit_cmp_and_jump_insns ( index, range, GTU, NULL_RTX, SImode, 1, oor_label ); -+ reg = gen_reg_rtx (SImode); -+ emit_move_insn ( reg, gen_rtx_LABEL_REF (VOIDmode, table_label)); -+ -+ if ( flag_pic ) -+ emit_jump_insn ( gen_tablejump_add ( reg, index, GEN_INT(4), table_label)); -+ else -+ emit_jump_insn ( -+ gen_tablejump_insn ( gen_rtx_MEM ( SImode, -+ gen_rtx_PLUS ( SImode, -+ reg, -+ gen_rtx_MULT ( SImode, -+ index, -+ GEN_INT(4)))), -+ table_label)); -+ DONE; -+ }" -+) -+ -+ -+ -+(define_insn "prefetch" -+ [(prefetch (match_operand:SI 0 "register_operand" "r") -+ (match_operand 1 "const_int_operand" "") -+ (match_operand 2 "const_int_operand" ""))] -+ "" -+ { -+ return "pref\t%0[0]"; -+ } -+ -+ [(set_attr "length" "4") -+ (set_attr "type" "load") -+ (set_attr "cc" "none")]) -+ -+ -+ -+;;============================================================================= -+;; prologue -+;;----------------------------------------------------------------------------- -+;; This pattern, if defined, emits RTL for entry to a function. The function -+;; entry i responsible for setting up the stack frame, initializing the frame -+;; pointer register, saving callee saved registers, etc. -+;;============================================================================= -+(define_expand "prologue" -+ [(clobber (const_int 0))] -+ "" -+ " -+ avr32_expand_prologue(); -+ DONE; -+ " -+ ) -+ -+;;============================================================================= -+;; eh_return -+;;----------------------------------------------------------------------------- -+;; This pattern, if defined, affects the way __builtin_eh_return, and -+;; thence the call frame exception handling library routines, are -+;; built. It is intended to handle non-trivial actions needed along -+;; the abnormal return path. -+;; -+;; The address of the exception handler to which the function should -+;; return is passed as operand to this pattern. It will normally need -+;; to copied by the pattern to some special register or memory -+;; location. If the pattern needs to determine the location of the -+;; target call frame in order to do so, it may use -+;; EH_RETURN_STACKADJ_RTX, if defined; it will have already been -+;; assigned. -+;; -+;; If this pattern is not defined, the default action will be to -+;; simply copy the return address to EH_RETURN_HANDLER_RTX. Either -+;; that macro or this pattern needs to be defined if call frame -+;; exception handling is to be used. -+ -+;; We can't expand this before we know where the link register is stored. -+(define_insn_and_split "eh_return" -+ [(unspec_volatile [(match_operand:SI 0 "register_operand" "r")] -+ VUNSPEC_EH_RETURN) -+ (clobber (match_scratch:SI 1 "=&r"))] -+ "" -+ "#" -+ "reload_completed" -+ [(const_int 0)] -+ " -+ { -+ avr32_set_return_address (operands[0], operands[1]); -+ DONE; -+ }" -+ ) -+ -+;;============================================================================= -+;; ffssi2 -+;;----------------------------------------------------------------------------- -+(define_insn "ffssi2" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (ffs:SI (match_operand:SI 1 "register_operand" "r"))) ] -+ "" -+ "mov %0, %1 -+ brev %0 -+ clz %0, %0 -+ sub %0, -1 -+ cp %0, 33 -+ moveq %0, 0" -+ [(set_attr "length" "18") -+ (set_attr "cc" "clobber")] -+ ) -+ -+ -+ -+;;============================================================================= -+;; swap_h -+;;----------------------------------------------------------------------------- -+(define_insn "*swap_h" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (ior:SI (ashift:SI (match_dup 0) (const_int 16)) -+ (lshiftrt:SI (match_dup 0) (const_int 16))))] -+ "" -+ "swap.h %0" -+ [(set_attr "length" "2")] -+ ) -+ -+(define_insn_and_split "bswap_16" -+ [ (set (match_operand:HI 0 "avr32_bswap_operand" "=r,RKs13,r") -+ (ior:HI (and:HI (lshiftrt:HI (match_operand:HI 1 "avr32_bswap_operand" "r,r,RKs13") -+ (const_int 8)) -+ (const_int 255)) -+ (ashift:HI (and:HI (match_dup 1) -+ (const_int 255)) -+ (const_int 8))))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ if ( REGNO(operands[0]) == REGNO(operands[1])) -+ return "swap.bh\t%0"; -+ else -+ return "mov\t%0, %1\;swap.bh\t%0"; -+ case 1: -+ return "stswp.h\t%0, %1"; -+ case 2: -+ return "ldswp.sh\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ -+ "(reload_completed && -+ REG_P(operands[0]) && REG_P(operands[1]) -+ && (REGNO(operands[0]) != REGNO(operands[1])))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 0) -+ (ior:HI (and:HI (lshiftrt:HI (match_dup 0) -+ (const_int 8)) -+ (const_int 255)) -+ (ashift:HI (and:HI (match_dup 0) -+ (const_int 255)) -+ (const_int 8))))] -+ "" -+ -+ [(set_attr "length" "4,4,4") -+ (set_attr "type" "alu,store,load_rm")] -+ ) -+ -+(define_insn_and_split "bswap_32" -+ [ (set (match_operand:SI 0 "avr32_bswap_operand" "=r,RKs14,r") -+ (ior:SI (ior:SI (lshiftrt:SI (and:SI (match_operand:SI 1 "avr32_bswap_operand" "=r,r,RKs14") -+ (const_int 4278190080)) -+ (const_int 24)) -+ (lshiftrt:SI (and:SI (match_dup 1) -+ (const_int 16711680)) -+ (const_int 8))) -+ (ior:SI (ashift:SI (and:SI (match_dup 1) -+ (const_int 65280)) -+ (const_int 8)) -+ (ashift:SI (and:SI (match_dup 1) -+ (const_int 255)) -+ (const_int 24)))))] -+ "" -+ { -+ switch ( which_alternative ){ -+ case 0: -+ if ( REGNO(operands[0]) == REGNO(operands[1])) -+ return "swap.b\t%0"; -+ else -+ return "mov\t%0, %1\;swap.b\t%0"; -+ case 1: -+ return "stswp.w\t%0, %1"; -+ case 2: -+ return "ldswp.w\t%0, %1"; -+ default: -+ abort(); -+ } -+ } -+ "(reload_completed && -+ REG_P(operands[0]) && REG_P(operands[1]) -+ && (REGNO(operands[0]) != REGNO(operands[1])))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 0) -+ (ior:SI (ior:SI (lshiftrt:SI (and:SI (match_dup 0) -+ (const_int 4278190080)) -+ (const_int 24)) -+ (lshiftrt:SI (and:SI (match_dup 0) -+ (const_int 16711680)) -+ (const_int 8))) -+ (ior:SI (ashift:SI (and:SI (match_dup 0) -+ (const_int 65280)) -+ (const_int 8)) -+ (ashift:SI (and:SI (match_dup 0) -+ (const_int 255)) -+ (const_int 24)))))] -+ "" -+ -+ [(set_attr "length" "4,4,4") -+ (set_attr "type" "alu,store,load_rm")] -+ ) -+ -+ -+;;============================================================================= -+;; blockage -+;;----------------------------------------------------------------------------- -+;; UNSPEC_VOLATILE is considered to use and clobber all hard registers and -+;; all of memory. This blocks insns from being moved across this point. -+ -+(define_insn "blockage" -+ [(unspec_volatile [(const_int 0)] VUNSPEC_BLOCKAGE)] -+ "" -+ "" -+ [(set_attr "length" "0")] -+) -+ -+;;============================================================================= -+;; clzsi2 -+;;----------------------------------------------------------------------------- -+(define_insn "clzsi2" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (clz:SI (match_operand:SI 1 "register_operand" "r"))) ] -+ "" -+ "clz %0, %1" -+ [(set_attr "length" "4") -+ (set_attr "cc" "set_z")] -+ ) -+ -+;;============================================================================= -+;; ctzsi2 -+;;----------------------------------------------------------------------------- -+(define_insn "ctzsi2" -+ [ (set (match_operand:SI 0 "register_operand" "=r,r") -+ (ctz:SI (match_operand:SI 1 "register_operand" "0,r"))) ] -+ "" -+ "@ -+ brev\t%0\;clz\t%0, %0 -+ mov\t%0, %1\;brev\t%0\;clz\t%0, %0" -+ [(set_attr "length" "8") -+ (set_attr "cc" "set_z")] -+ ) -+ -+;;============================================================================= -+;; cache instructions -+;;----------------------------------------------------------------------------- -+(define_insn "cache" -+ [ (unspec_volatile [(match_operand:SI 0 "register_operand" "r") -+ (match_operand:SI 1 "immediate_operand" "Ku05")] VUNSPEC_CACHE)] -+ "" -+ "cache %0[0], %1" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "sync" -+ [ (unspec_volatile [(match_operand:SI 0 "immediate_operand" "Ku08")] VUNSPEC_SYNC)] -+ "" -+ "sync %0" -+ [(set_attr "length" "4")] -+ ) -+ -+;;============================================================================= -+;; TLB instructions -+;;----------------------------------------------------------------------------- -+(define_insn "tlbr" -+ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBR)] -+ "" -+ "tlbr" -+ [(set_attr "length" "2")] -+ ) -+ -+(define_insn "tlbw" -+ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBW)] -+ "" -+ "tlbw" -+ [(set_attr "length" "2")] -+ ) -+ -+(define_insn "tlbs" -+ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBS)] -+ "" -+ "tlbs" -+ [(set_attr "length" "2")] -+ ) -+ -+;;============================================================================= -+;; Breakpoint instruction -+;;----------------------------------------------------------------------------- -+(define_insn "breakpoint" -+ [ (unspec_volatile [(const_int 0)] VUNSPEC_BREAKPOINT)] -+ "" -+ "breakpoint" -+ [(set_attr "length" "2")] -+ ) -+ -+ -+;;============================================================================= -+;; mtsr/mfsr instruction -+;;----------------------------------------------------------------------------- -+(define_insn "mtsr" -+ [ (unspec_volatile [(match_operand 0 "immediate_operand" "i") -+ (match_operand:SI 1 "register_operand" "r")] VUNSPEC_MTSR)] -+ "" -+ "mtsr\t%0, %1" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mfsr" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "i")] VUNSPEC_MFSR)) ] -+ "" -+ "mfsr\t%0, %1" -+ [(set_attr "length" "4")] -+ ) -+ -+;;============================================================================= -+;; mtdr/mfdr instruction -+;;----------------------------------------------------------------------------- -+(define_insn "mtdr" -+ [ (unspec_volatile [(match_operand 0 "immediate_operand" "i") -+ (match_operand:SI 1 "register_operand" "r")] VUNSPEC_MTDR)] -+ "" -+ "mtdr\t%0, %1" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mfdr" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "i")] VUNSPEC_MFDR)) ] -+ "" -+ "mfdr\t%0, %1" -+ [(set_attr "length" "4")] -+ ) -+ -+;;============================================================================= -+;; musfr -+;;----------------------------------------------------------------------------- -+(define_insn "musfr" -+ [ (unspec_volatile [(match_operand:SI 0 "register_operand" "r")] VUNSPEC_MUSFR)] -+ "" -+ "musfr\t%0" -+ [(set_attr "length" "2") -+ (set_attr "cc" "clobber")] -+ ) -+ -+(define_insn "mustr" -+ [ (set (match_operand:SI 0 "register_operand" "=r") -+ (unspec_volatile:SI [(const_int 0)] VUNSPEC_MUSTR)) ] -+ "" -+ "mustr\t%0" -+ [(set_attr "length" "2")] -+ ) -+ -+;;============================================================================= -+;; Flush Return Stack instruction -+;;----------------------------------------------------------------------------- -+(define_insn "frs" -+ [ (unspec_volatile [(const_int 0)] VUNSPEC_FRS)] -+ "" -+ "frs" -+ [(set_attr "length" "2") -+ (set_attr "cc" "none")] -+ ) -+ -+ -+;;============================================================================= -+;; Saturation Round Scale instruction -+;;----------------------------------------------------------------------------- -+(define_insn "sats" -+ [ (set (match_operand:SI 0 "register_operand" "+r") -+ (unspec:SI [(match_dup 0) -+ (match_operand 1 "immediate_operand" "Ku05") -+ (match_operand 2 "immediate_operand" "Ku05")] -+ UNSPEC_SATS)) ] -+ "TARGET_DSP" -+ "sats\t%0 >> %1, %2" -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4")] -+ ) -+ -+(define_insn "satu" -+ [ (set (match_operand:SI 0 "register_operand" "+r") -+ (unspec:SI [(match_dup 0) -+ (match_operand 1 "immediate_operand" "Ku05") -+ (match_operand 2 "immediate_operand" "Ku05")] -+ UNSPEC_SATU)) ] -+ "TARGET_DSP" -+ "satu\t%0 >> %1, %2" -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4")] -+ ) -+ -+(define_insn "satrnds" -+ [ (set (match_operand:SI 0 "register_operand" "+r") -+ (unspec:SI [(match_dup 0) -+ (match_operand 1 "immediate_operand" "Ku05") -+ (match_operand 2 "immediate_operand" "Ku05")] -+ UNSPEC_SATRNDS)) ] -+ "TARGET_DSP" -+ "satrnds\t%0 >> %1, %2" -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4")] -+ ) -+ -+(define_insn "satrndu" -+ [ (set (match_operand:SI 0 "register_operand" "+r") -+ (unspec:SI [(match_dup 0) -+ (match_operand 1 "immediate_operand" "Ku05") -+ (match_operand 2 "immediate_operand" "Ku05")] -+ UNSPEC_SATRNDU)) ] -+ "TARGET_DSP" -+ "sats\t%0 >> %1, %2" -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4")] -+ ) -+ -+;; Special patterns for dealing with the constant pool -+ -+(define_insn "align_4" -+ [(unspec_volatile [(const_int 0)] VUNSPEC_ALIGN)] -+ "" -+ { -+ assemble_align (32); -+ return ""; -+ } -+ [(set_attr "length" "2")] -+) -+ -+(define_insn "consttable_start" -+ [(unspec_volatile [(const_int 0)] VUNSPEC_POOL_START)] -+ "" -+ { -+ return ".cpool"; -+ } -+ [(set_attr "length" "0")] -+ ) -+ -+(define_insn "consttable_end" -+ [(unspec_volatile [(const_int 0)] VUNSPEC_POOL_END)] -+ "" -+ { -+ making_const_table = FALSE; -+ return ""; -+ } -+ [(set_attr "length" "0")] -+) -+ -+ -+(define_insn "consttable_4" -+ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_4)] -+ "" -+ { -+ making_const_table = TRUE; -+ switch (GET_MODE_CLASS (GET_MODE (operands[0]))) -+ { -+ case MODE_FLOAT: -+ { -+ REAL_VALUE_TYPE r; -+ char real_string[1024]; -+ REAL_VALUE_FROM_CONST_DOUBLE (r, operands[0]); -+ real_to_decimal(real_string, &r, 1024, 0, 1); -+ asm_fprintf (asm_out_file, "\t.float\t%s\n", real_string); -+ break; -+ } -+ default: -+ assemble_integer (operands[0], 4, 0, 1); -+ break; -+ } -+ return ""; -+ } -+ [(set_attr "length" "4")] -+) -+ -+(define_insn "consttable_8" -+ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_8)] -+ "" -+ { -+ making_const_table = TRUE; -+ switch (GET_MODE_CLASS (GET_MODE (operands[0]))) -+ { -+ case MODE_FLOAT: -+ { -+ REAL_VALUE_TYPE r; -+ char real_string[1024]; -+ REAL_VALUE_FROM_CONST_DOUBLE (r, operands[0]); -+ real_to_decimal(real_string, &r, 1024, 0, 1); -+ asm_fprintf (asm_out_file, "\t.double\t%s\n", real_string); -+ break; -+ } -+ default: -+ assemble_integer(operands[0], 8, 0, 1); -+ break; -+ } -+ return ""; -+ } -+ [(set_attr "length" "8")] -+) -+ -+(define_insn "consttable_16" -+ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_16)] -+ "" -+ { -+ making_const_table = TRUE; -+ assemble_integer(operands[0], 16, 0, 1); -+ return ""; -+ } -+ [(set_attr "length" "16")] -+) -+ -+;;============================================================================= -+;; coprocessor instructions -+;;----------------------------------------------------------------------------- -+(define_insn "cop" -+ [ (unspec_volatile [(match_operand 0 "immediate_operand" "Ku03") -+ (match_operand 1 "immediate_operand" "Ku04") -+ (match_operand 2 "immediate_operand" "Ku04") -+ (match_operand 3 "immediate_operand" "Ku04") -+ (match_operand 4 "immediate_operand" "Ku07")] VUNSPEC_COP)] -+ "" -+ "cop\tcp%0, cr%1, cr%2, cr%3, %4" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mvcrsi" -+ [ (set (match_operand:SI 0 "avr32_cop_move_operand" "=r,<,Z") -+ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "Ku03,Ku03,Ku03") -+ (match_operand 2 "immediate_operand" "Ku04,Ku04,Ku04")] -+ VUNSPEC_MVCR)) ] -+ "" -+ "@ -+ mvcr.w\tcp%1, %0, cr%2 -+ stcm.w\tcp%1, %0, cr%2 -+ stc.w\tcp%1, %0, cr%2" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mvcrdi" -+ [ (set (match_operand:DI 0 "avr32_cop_move_operand" "=r,<,Z") -+ (unspec_volatile:DI [(match_operand 1 "immediate_operand" "Ku03,Ku03,Ku03") -+ (match_operand 2 "immediate_operand" "Ku04,Ku04,Ku04")] -+ VUNSPEC_MVCR)) ] -+ "" -+ "@ -+ mvcr.d\tcp%1, %0, cr%2 -+ stcm.d\tcp%1, %0, cr%2-cr%i2 -+ stc.d\tcp%1, %0, cr%2" -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mvrcsi" -+ [ (unspec_volatile:SI [(match_operand 0 "immediate_operand" "Ku03,Ku03,Ku03") -+ (match_operand 1 "immediate_operand" "Ku04,Ku04,Ku04") -+ (match_operand:SI 2 "avr32_cop_move_operand" "r,>,Z")] -+ VUNSPEC_MVRC)] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "mvrc.w\tcp%0, cr%1, %2"; -+ case 1: -+ return "ldcm.w\tcp%0, %2, cr%1"; -+ case 2: -+ return "ldc.w\tcp%0, cr%1, %2"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "4")] -+ ) -+ -+(define_insn "mvrcdi" -+ [ (unspec_volatile:DI [(match_operand 0 "immediate_operand" "Ku03,Ku03,Ku03") -+ (match_operand 1 "immediate_operand" "Ku04,Ku04,Ku04") -+ (match_operand:DI 2 "avr32_cop_move_operand" "r,>,Z")] -+ VUNSPEC_MVRC)] -+ "" -+ { -+ switch (which_alternative){ -+ case 0: -+ return "mvrc.d\tcp%0, cr%1, %2"; -+ case 1: -+ return "ldcm.d\tcp%0, %2, cr%1-cr%i1"; -+ case 2: -+ return "ldc.d\tcp%0, cr%1, %2"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "length" "4")] -+ ) -+ -+;;============================================================================= -+;; epilogue -+;;----------------------------------------------------------------------------- -+;; This pattern emits RTL for exit from a function. The function exit is -+;; responsible for deallocating the stack frame, restoring callee saved -+;; registers and emitting the return instruction. -+;; ToDo: using TARGET_ASM_FUNCTION_PROLOGUE instead. -+;;============================================================================= -+(define_expand "epilogue" -+ [(unspec_volatile [(return)] VUNSPEC_EPILOGUE)] -+ "" -+ " -+ if (USE_RETURN_INSN (FALSE)){ -+ emit_jump_insn (gen_return ()); -+ DONE; -+ } -+ emit_jump_insn (gen_rtx_UNSPEC_VOLATILE (VOIDmode, -+ gen_rtvec (1, -+ gen_rtx_RETURN (VOIDmode)), -+ VUNSPEC_EPILOGUE)); -+ DONE; -+ " -+ ) -+ -+(define_insn "*epilogue_insns" -+ [(unspec_volatile [(return)] VUNSPEC_EPILOGUE)] -+ "" -+ { -+ avr32_output_return_instruction (FALSE, FALSE, NULL, NULL); -+ return ""; -+ } -+ ; Length is absolute worst case -+ [(set_attr "type" "branch") -+ (set_attr "length" "12")] -+ ) -+ -+(define_insn "*epilogue_insns_ret_imm" -+ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) -+ (use (reg RETVAL_REGNUM)) -+ (unspec_volatile [(return)] VUNSPEC_EPILOGUE)])] -+ "((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" -+ { -+ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[0]); -+ return ""; -+ } -+ ; Length is absolute worst case -+ [(set_attr "type" "branch") -+ (set_attr "length" "12")] -+ ) -+ -+(define_insn "sibcall_epilogue" -+ [(unspec_volatile [(const_int 0)] VUNSPEC_EPILOGUE)] -+ "" -+ { -+ avr32_output_return_instruction (FALSE, FALSE, NULL, NULL); -+ return ""; -+ } -+;; Length is absolute worst case -+ [(set_attr "type" "branch") -+ (set_attr "length" "12")] -+ ) -+ -+(define_insn "*sibcall_epilogue_insns_ret_imm" -+ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) -+ (use (reg RETVAL_REGNUM)) -+ (unspec_volatile [(const_int 0)] VUNSPEC_EPILOGUE)])] -+ "((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" -+ { -+ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[0]); -+ return ""; -+ } -+ ; Length is absolute worst case -+ [(set_attr "type" "branch") -+ (set_attr "length" "12")] -+ ) -+ -+(define_insn "ldxi" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (mem:SI (plus:SI -+ (match_operand:SI 1 "register_operand" "r") -+ (mult:SI (zero_extract:SI (match_operand:SI 2 "register_operand" "r") -+ (const_int 8) -+ (match_operand:SI 3 "immediate_operand" "Ku05")) -+ (const_int 4)))))] -+ "(INTVAL(operands[3]) == 24 || INTVAL(operands[3]) == 16 || INTVAL(operands[3]) == 8 -+ || INTVAL(operands[3]) == 0)" -+ { -+ switch ( INTVAL(operands[3]) ){ -+ case 0: -+ return "ld.w %0, %1[%2:b << 2]"; -+ case 8: -+ return "ld.w %0, %1[%2:l << 2]"; -+ case 16: -+ return "ld.w %0, %1[%2:u << 2]"; -+ case 24: -+ return "ld.w %0, %1[%2:t << 2]"; -+ default: -+ internal_error("illegal operand for ldxi"); -+ } -+ } -+ [(set_attr "type" "load") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -+ -+ -+ -+ -+ -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; sub r8, r7, 8 -+;; st.w r8[0x0], r12 -+;; to -+;; sub r8, r7, 8 -+;; st.w r7[-0x8], r12 -+;;============================================================================= -+; (set (reg:SI 9 r8) -+; (plus:SI (reg/f:SI 6 r7) -+; (const_int ...))) -+; (set (mem:SI (reg:SI 9 r8)) -+; (reg:SI 12 r12)) -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (plus:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" ""))) -+ (set (mem:SI (match_dup 0)) -+ (match_operand:SI 3 "register_operand" ""))] -+ "REGNO(operands[0]) != REGNO(operands[1]) && avr32_const_ok_for_constraint_p(INTVAL(operands[2]), 'K', \"Ks16\")" -+ [(set (match_dup 0) -+ (plus:SI (match_dup 1) -+ (match_dup 2))) -+ (set (mem:SI (plus:SI (match_dup 1) -+ (match_dup 2))) -+ (match_dup 3))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; sub r6, r7, 4 -+;; ld.w r6, r6[0x0] -+;; to -+;; sub r6, r7, 4 -+;; ld.w r6, r7[-0x4] -+;;============================================================================= -+; (set (reg:SI 7 r6) -+; (plus:SI (reg/f:SI 6 r7) -+; (const_int -4 [0xfffffffc]))) -+; (set (reg:SI 7 r6) -+; (mem:SI (reg:SI 7 r6))) -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (plus:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "immediate_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (mem:SI (match_dup 0)))] -+ "REGNO(operands[0]) != REGNO(operands[1]) && avr32_const_ok_for_constraint_p(INTVAL(operands[2]), 'K', \"Ks16\")" -+ [(set (match_dup 0) -+ (plus:SI (match_dup 1) -+ (match_dup 2))) -+ (set (match_dup 3) -+ (mem:SI (plus:SI (match_dup 1) -+ (match_dup 2))))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; ld.sb r0, r7[-0x6] -+;; cashs.b r0 -+;; to -+;; ld.sb r0, r7[-0x6] -+;;============================================================================= -+(define_peephole2 -+ [(set (match_operand:QI 0 "register_operand" "") -+ (match_operand:QI 1 "load_sb_memory_operand" "")) -+ (set (match_operand:SI 2 "register_operand" "") -+ (sign_extend:SI (match_dup 0)))] -+ "(REGNO(operands[0]) == REGNO(operands[2]) || peep2_reg_dead_p(2, operands[0]))" -+ [(set (match_dup 2) -+ (sign_extend:SI (match_dup 1)))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; ld.ub r0, r7[-0x6] -+;; cashu.b r0 -+;; to -+;; ld.ub r0, r7[-0x6] -+;;============================================================================= -+(define_peephole2 -+ [(set (match_operand:QI 0 "register_operand" "") -+ (match_operand:QI 1 "memory_operand" "")) -+ (set (match_operand:SI 2 "register_operand" "") -+ (zero_extend:SI (match_dup 0)))] -+ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 2) -+ (zero_extend:SI (match_dup 1)))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; ld.sh r0, r7[-0x6] -+;; casts.h r0 -+;; to -+;; ld.sh r0, r7[-0x6] -+;;============================================================================= -+(define_peephole2 -+ [(set (match_operand:HI 0 "register_operand" "") -+ (match_operand:HI 1 "memory_operand" "")) -+ (set (match_operand:SI 2 "register_operand" "") -+ (sign_extend:SI (match_dup 0)))] -+ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 2) -+ (sign_extend:SI (match_dup 1)))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; ld.uh r0, r7[-0x6] -+;; castu.h r0 -+;; to -+;; ld.uh r0, r7[-0x6] -+;;============================================================================= -+(define_peephole2 -+ [(set (match_operand:HI 0 "register_operand" "") -+ (match_operand:HI 1 "memory_operand" "")) -+ (set (match_operand:SI 2 "register_operand" "") -+ (zero_extend:SI (match_dup 0)))] -+ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 2) -+ (zero_extend:SI (match_dup 1)))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; mul rd, rx, ry -+;; add rd2, rd -+;; to -+;; mac rd2, rx, ry -+;;============================================================================= -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (mult:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "register_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (plus:SI (match_dup 3) -+ (match_dup 0)))] -+ "peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 3) -+ (plus:SI (mult:SI (match_dup 1) -+ (match_dup 2)) -+ (match_dup 3)))] -+ "") -+ -+ -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Changing -+;; bfextu rd, rs, k5, 1 or and(h/l) rd, one_bit_set_mask -+;; to -+;; bld rs, k5 -+;; -+;; If rd is dead after the operation. -+;;============================================================================= -+(define_peephole2 -+ [ (set (match_operand:SI 0 "register_operand" "") -+ (zero_extract:SI (match_operand:SI 1 "register_operand" "") -+ (const_int 1) -+ (match_operand:SI 2 "immediate_operand" ""))) -+ (set (cc0) -+ (match_dup 0))] -+ "peep2_reg_dead_p(2, operands[0])" -+ [(set (cc0) -+ (and:SI (match_dup 1) -+ (match_dup 2)))] -+ "operands[2] = GEN_INT(1 << INTVAL(operands[2]));") -+ -+(define_peephole2 -+ [ (set (match_operand:SI 0 "register_operand" "") -+ (and:SI (match_operand:SI 1 "register_operand" "") -+ (match_operand:SI 2 "one_bit_set_operand" ""))) -+ (set (cc0) -+ (match_dup 0))] -+ "peep2_reg_dead_p(2, operands[0])" -+ [(set (cc0) -+ (and:SI (match_dup 1) -+ (match_dup 2)))] -+ "") -+ -+;;============================================================================= -+;; Peephole optimizing -+;;----------------------------------------------------------------------------- -+;; Load with extracted index: ld.w Rd, Rb[Ri:{t/u/b/l} << 2] -+;; -+;;============================================================================= -+ -+ -+(define_peephole -+ [(set (match_operand:SI 0 "register_operand" "") -+ (zero_extract:SI (match_operand:SI 1 "register_operand" "") -+ (const_int 8) -+ (match_operand:SI 2 "avr32_extract_shift_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 4 "register_operand" ""))))] -+ -+ "(dead_or_set_p(insn, operands[0]))" -+ { -+ switch ( INTVAL(operands[2]) ){ -+ case 0: -+ return "ld.w %3, %4[%1:b << 2]"; -+ case 8: -+ return "ld.w %3, %4[%1:l << 2]"; -+ case 16: -+ return "ld.w %3, %4[%1:u << 2]"; -+ case 24: -+ return "ld.w %3, %4[%1:t << 2]"; -+ default: -+ internal_error("illegal operand for ldxi"); -+ } -+ } -+ [(set_attr "type" "load") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")] -+ ) -+ -+ -+ -+(define_peephole -+ [(set (match_operand:SI 0 "register_operand" "") -+ (and:SI (match_operand:SI 1 "register_operand" "") (const_int 255))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 3 "register_operand" ""))))] -+ -+ "(dead_or_set_p(insn, operands[0]))" -+ -+ "ld.w %2, %3[%1:b << 2]" -+ [(set_attr "type" "load") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")] -+ ) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (zero_extract:SI (match_operand:SI 1 "register_operand" "") -+ (const_int 8) -+ (match_operand:SI 2 "avr32_extract_shift_operand" ""))) -+ (set (match_operand:SI 3 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 4 "register_operand" ""))))] -+ -+ "(peep2_reg_dead_p(2, operands[0])) -+ || (REGNO(operands[0]) == REGNO(operands[3]))" -+ [(set (match_dup 3) -+ (mem:SI (plus:SI -+ (match_dup 4) -+ (mult:SI (zero_extract:SI (match_dup 1) -+ (const_int 8) -+ (match_dup 2)) -+ (const_int 4)))))] -+ ) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (zero_extend:SI (match_operand:QI 1 "register_operand" ""))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 3 "register_operand" ""))))] -+ -+ "(peep2_reg_dead_p(2, operands[0])) -+ || (REGNO(operands[0]) == REGNO(operands[2]))" -+ [(set (match_dup 2) -+ (mem:SI (plus:SI -+ (match_dup 3) -+ (mult:SI (zero_extract:SI (match_dup 1) -+ (const_int 8) -+ (const_int 0)) -+ (const_int 4)))))] -+ "operands[1] = gen_rtx_REG(SImode, REGNO(operands[1]));" -+ ) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (and:SI (match_operand:SI 1 "register_operand" "") -+ (const_int 255))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 3 "register_operand" ""))))] -+ -+ "(peep2_reg_dead_p(2, operands[0])) -+ || (REGNO(operands[0]) == REGNO(operands[2]))" -+ [(set (match_dup 2) -+ (mem:SI (plus:SI -+ (match_dup 3) -+ (mult:SI (zero_extract:SI (match_dup 1) -+ (const_int 8) -+ (const_int 0)) -+ (const_int 4)))))] -+ "" -+ ) -+ -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (lshiftrt:SI (match_operand:SI 1 "register_operand" "") -+ (const_int 24))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) -+ (match_operand:SI 3 "register_operand" ""))))] -+ -+ "(peep2_reg_dead_p(2, operands[0])) -+ || (REGNO(operands[0]) == REGNO(operands[2]))" -+ [(set (match_dup 2) -+ (mem:SI (plus:SI -+ (match_dup 3) -+ (mult:SI (zero_extract:SI (match_dup 1) -+ (const_int 8) -+ (const_int 24)) -+ (const_int 4)))))] -+ "" -+ ) -+ -+ -+;;************************************************ -+;; ANDN -+;; -+;;************************************************ -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (not:SI (match_operand:SI 1 "register_operand" ""))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (and:SI (match_dup 2) -+ (match_dup 0)))] -+ "peep2_reg_dead_p(2, operands[0])" -+ -+ [(set (match_dup 2) -+ (and:SI (match_dup 2) -+ (not:SI (match_dup 1)) -+ ))] -+ "" -+) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (not:SI (match_operand:SI 1 "register_operand" ""))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (and:SI (match_dup 0) -+ (match_dup 2) -+ ))] -+ "peep2_reg_dead_p(2, operands[0])" -+ -+ [(set (match_dup 2) -+ (and:SI (match_dup 2) -+ (not:SI (match_dup 1)) -+ ))] -+ -+ "" -+) -+ -+ -+;;================================================================= -+;; Addabs peephole -+;;================================================================= -+ -+(define_peephole -+ [(set (match_operand:SI 2 "register_operand" "=r") -+ (abs:SI (match_operand:SI 1 "register_operand" "r"))) -+ (set (match_operand:SI 0 "register_operand" "=r") -+ (plus:SI (match_operand:SI 3 "register_operand" "r") -+ (match_dup 2)))] -+ "dead_or_set_p(insn, operands[2])" -+ "addabs %0, %3, %1" -+ [(set_attr "length" "4") -+ (set_attr "cc" "set_z")]) -+ -+(define_peephole -+ [(set (match_operand:SI 2 "register_operand" "=r") -+ (abs:SI (match_operand:SI 1 "register_operand" "r"))) -+ (set (match_operand:SI 0 "register_operand" "=r") -+ (plus:SI (match_dup 2) -+ (match_operand:SI 3 "register_operand" "r")))] -+ "dead_or_set_p(insn, operands[2])" -+ "addabs %0, %3, %1" -+ [(set_attr "length" "4") -+ (set_attr "cc" "set_z")]) -+ -+ -+;;================================================================= -+;; Detect roundings -+;;================================================================= -+ -+(define_insn "*round" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (ashiftrt:SI (plus:SI (match_operand:SI 1 "register_operand" "0") -+ (match_operand:SI 2 "immediate_operand" "i")) -+ (match_operand:SI 3 "immediate_operand" "i")))] -+ "avr32_rnd_operands(operands[2], operands[3])" -+ -+ "satrnds %0 >> %3, 31" -+ -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4")] -+ -+ ) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (plus:SI (match_dup 0) -+ (match_operand:SI 1 "immediate_operand" ""))) -+ (set (match_dup 0) -+ (ashiftrt:SI (match_dup 0) -+ (match_operand:SI 2 "immediate_operand" "")))] -+ "avr32_rnd_operands(operands[1], operands[2])" -+ -+ [(set (match_dup 0) -+ (ashiftrt:SI (plus:SI (match_dup 0) -+ (match_dup 1)) -+ (match_dup 2)))] -+ ) -+ -+(define_peephole -+ [(set (match_operand:SI 0 "register_operand" "r") -+ (plus:SI (match_dup 0) -+ (match_operand:SI 1 "immediate_operand" "i"))) -+ (set (match_dup 0) -+ (ashiftrt:SI (match_dup 0) -+ (match_operand:SI 2 "immediate_operand" "i")))] -+ "avr32_rnd_operands(operands[1], operands[2])" -+ -+ "satrnds %0 >> %2, 31" -+ -+ [(set_attr "type" "alu_sat") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")] -+ -+ ) -+ -+ -+;;================================================================= -+;; mcall -+;;================================================================= -+(define_peephole -+ [(set (match_operand:SI 0 "register_operand" "") -+ (match_operand 1 "avr32_const_pool_ref_operand" "")) -+ (parallel [(call (mem:SI (match_dup 0)) -+ (match_operand 2 "" "")) -+ (clobber (reg:SI LR_REGNUM))])] -+ "dead_or_set_p(insn, operands[0])" -+ "mcall %1" -+ [(set_attr "type" "call") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")] -+) -+ -+(define_peephole -+ [(set (match_operand:SI 2 "register_operand" "") -+ (match_operand 1 "avr32_const_pool_ref_operand" "")) -+ (parallel [(set (match_operand 0 "register_operand" "") -+ (call (mem:SI (match_dup 2)) -+ (match_operand 3 "" ""))) -+ (clobber (reg:SI LR_REGNUM))])] -+ "dead_or_set_p(insn, operands[2])" -+ "mcall %1" -+ [(set_attr "type" "call") -+ (set_attr "length" "4") -+ (set_attr "cc" "call_set")] -+) -+ -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (match_operand 1 "avr32_const_pool_ref_operand" "")) -+ (parallel [(call (mem:SI (match_dup 0)) -+ (match_operand 2 "" "")) -+ (clobber (reg:SI LR_REGNUM))])] -+ "peep2_reg_dead_p(2, operands[0])" -+ [(parallel [(call (mem:SI (match_dup 1)) -+ (match_dup 2)) -+ (clobber (reg:SI LR_REGNUM))])] -+ "" -+) -+ -+(define_peephole2 -+ [(set (match_operand:SI 0 "register_operand" "") -+ (match_operand 1 "avr32_const_pool_ref_operand" "")) -+ (parallel [(set (match_operand 2 "register_operand" "") -+ (call (mem:SI (match_dup 0)) -+ (match_operand 3 "" ""))) -+ (clobber (reg:SI LR_REGNUM))])] -+ "(peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[2]) == REGNO(operands[0])))" -+ [(parallel [(set (match_dup 2) -+ (call (mem:SI (match_dup 1)) -+ (match_dup 3))) -+ (clobber (reg:SI LR_REGNUM))])] -+ "" -+) -+ -+;;================================================================= -+;; Returning a value -+;;================================================================= -+ -+ -+(define_peephole -+ [(set (match_operand 0 "register_operand" "") -+ (match_operand 1 "register_operand" "")) -+ (return)] -+ "USE_RETURN_INSN (TRUE) && (REGNO(operands[0]) == RETVAL_REGNUM) -+ && (REGNO(operands[1]) != LR_REGNUM) -+ && (REGNO_REG_CLASS(REGNO(operands[1])) == GENERAL_REGS)" -+ "retal %1" -+ [(set_attr "type" "call") -+ (set_attr "length" "2")] -+ ) -+ -+ -+(define_peephole -+ [(set (match_operand 0 "register_operand" "r") -+ (match_operand 1 "immediate_operand" "i")) -+ (return)] -+ "(USE_RETURN_INSN (FALSE) && (REGNO(operands[0]) == RETVAL_REGNUM) && -+ ((INTVAL(operands[1]) == -1) || (INTVAL(operands[1]) == 0) || (INTVAL(operands[1]) == 1)))" -+ { -+ avr32_output_return_instruction (TRUE, FALSE, NULL, operands[1]); -+ return ""; -+ } -+ [(set_attr "type" "call") -+ (set_attr "length" "4")] -+ ) -+ -+(define_peephole -+ [(set (match_operand 0 "register_operand" "r") -+ (match_operand 1 "immediate_operand" "i")) -+ (unspec_volatile [(return)] VUNSPEC_EPILOGUE)] -+ "(REGNO(operands[0]) == RETVAL_REGNUM) && -+ ((INTVAL(operands[1]) == -1) || (INTVAL(operands[1]) == 0) || (INTVAL(operands[1]) == 1))" -+ { -+ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[1]); -+ return ""; -+ } -+ ; Length is absolute worst case -+ [(set_attr "type" "branch") -+ (set_attr "length" "12")] -+ ) -+ -+(define_peephole -+ [(set (match_operand 0 "register_operand" "=r") -+ (if_then_else (match_operator 1 "avr32_comparison_operator" -+ [(match_operand 4 "register_operand" "r") -+ (match_operand 5 "register_immediate_operand" "rKs21")]) -+ (match_operand 2 "avr32_cond_register_immediate_operand" "rKs08") -+ (match_operand 3 "avr32_cond_register_immediate_operand" "rKs08"))) -+ (return)] -+ "USE_RETURN_INSN (TRUE) && (REGNO(operands[0]) == RETVAL_REGNUM)" -+ { -+ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); -+ -+ if ( GET_CODE(operands[2]) == REG -+ && GET_CODE(operands[3]) == REG -+ && REGNO(operands[2]) != LR_REGNUM -+ && REGNO(operands[3]) != LR_REGNUM ){ -+ return "ret%1 %2\;ret%i1 %3"; -+ } else if ( GET_CODE(operands[2]) == REG -+ && GET_CODE(operands[3]) == CONST_INT ){ -+ if ( INTVAL(operands[3]) == -1 -+ || INTVAL(operands[3]) == 0 -+ || INTVAL(operands[3]) == 1 ){ -+ return "ret%1 %2\;ret%i1 %d3"; -+ } else { -+ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; -+ } -+ } else if ( GET_CODE(operands[2]) == CONST_INT -+ && GET_CODE(operands[3]) == REG ){ -+ if ( INTVAL(operands[2]) == -1 -+ || INTVAL(operands[2]) == 0 -+ || INTVAL(operands[2]) == 1 ){ -+ return "ret%1 %d2\;ret%i1 %3"; -+ } else { -+ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; -+ } -+ } else { -+ if ( (INTVAL(operands[2]) == -1 -+ || INTVAL(operands[2]) == 0 -+ || INTVAL(operands[2]) == 1 ) -+ && (INTVAL(operands[3]) == -1 -+ || INTVAL(operands[3]) == 0 -+ || INTVAL(operands[3]) == 1 )){ -+ return "ret%1 %d2\;ret%i1 %d3"; -+ } else { -+ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; -+ } -+ } -+ } -+ -+ [(set_attr "length" "10") -+ (set_attr "cc" "none") -+ (set_attr "type" "call")]) -+ -+ -+;;================================================================= -+;; mulnhh.w -+;;================================================================= -+ -+(define_peephole2 -+ [(set (match_operand:HI 0 "register_operand" "") -+ (neg:HI (match_operand:HI 1 "register_operand" ""))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mult:SI -+ (sign_extend:SI (match_dup 0)) -+ (sign_extend:SI (match_operand:HI 3 "register_operand" ""))))] -+ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[2]) == REGNO(operands[0]))" -+ [ (set (match_dup 2) -+ (mult:SI -+ (sign_extend:SI (neg:HI (match_dup 1))) -+ (sign_extend:SI (match_dup 3))))] -+ "" -+ ) -+ -+(define_peephole2 -+ [(set (match_operand:HI 0 "register_operand" "") -+ (neg:HI (match_operand:HI 1 "register_operand" ""))) -+ (set (match_operand:SI 2 "register_operand" "") -+ (mult:SI -+ (sign_extend:SI (match_operand:HI 3 "register_operand" "")) -+ (sign_extend:SI (match_dup 0))))] -+ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[2]) == REGNO(operands[0]))" -+ [ (set (match_dup 2) -+ (mult:SI -+ (sign_extend:SI (neg:HI (match_dup 1))) -+ (sign_extend:SI (match_dup 3))))] -+ "" -+ ) -+ -+ -+ -+;;================================================================= -+;; Vector set and extract operations -+;;================================================================= -+(define_insn "vec_setv2hi_hi" -+ [(set (match_operand:V2HI 0 "register_operand" "=r") -+ (vec_merge:V2HI -+ (match_dup 0) -+ (vec_duplicate:V2HI -+ (match_operand:HI 1 "register_operand" "r")) -+ (const_int 1)))] -+ "" -+ "bfins\t%0, %1, 16, 16" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")]) -+ -+(define_insn "vec_setv2hi_lo" -+ [(set (match_operand:V2HI 0 "register_operand" "+r") -+ (vec_merge:V2HI -+ (match_dup 0) -+ (vec_duplicate:V2HI -+ (match_operand:HI 1 "register_operand" "r")) -+ (const_int 2)))] -+ "" -+ "bfins\t%0, %1, 0, 16" -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")]) -+ -+(define_expand "vec_setv2hi" -+ [(set (match_operand:V2HI 0 "register_operand" "") -+ (vec_merge:V2HI -+ (match_dup 0) -+ (vec_duplicate:V2HI -+ (match_operand:HI 1 "register_operand" "")) -+ (match_operand 2 "immediate_operand" "")))] -+ "" -+ { operands[2] = GEN_INT(INTVAL(operands[2]) + 1); } -+ ) -+ -+(define_insn "vec_extractv2hi" -+ [(set (match_operand:HI 0 "register_operand" "=r") -+ (vec_select:HI -+ (match_operand:V2HI 1 "register_operand" "r") -+ (parallel [(match_operand:SI 2 "immediate_operand" "i")])))] -+ "" -+ { -+ if ( INTVAL(operands[2]) == 0 ) -+ return "bfextu\t%0, %1, 16, 16"; -+ else -+ return "bfextu\t%0, %1, 0, 16"; -+ } -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")]) -+ -+(define_insn "vec_extractv4qi" -+ [(set (match_operand:QI 0 "register_operand" "=r") -+ (vec_select:QI -+ (match_operand:V4QI 1 "register_operand" "r") -+ (parallel [(match_operand:SI 2 "immediate_operand" "i")])))] -+ "" -+ { -+ switch ( INTVAL(operands[2]) ){ -+ case 0: -+ return "bfextu\t%0, %1, 24, 8"; -+ case 1: -+ return "bfextu\t%0, %1, 16, 8"; -+ case 2: -+ return "bfextu\t%0, %1, 8, 8"; -+ case 3: -+ return "bfextu\t%0, %1, 0, 8"; -+ default: -+ abort(); -+ } -+ } -+ [(set_attr "type" "alu") -+ (set_attr "length" "4") -+ (set_attr "cc" "clobber")]) -+ -+ -+(define_insn "concatv2hi" -+ [(set (match_operand:V2HI 0 "register_operand" "=r, r, r") -+ (vec_concat:V2HI -+ (match_operand:HI 1 "register_operand" "r, r, 0") -+ (match_operand:HI 2 "register_operand" "r, 0, r")))] -+ "" -+ "@ -+ mov\t%0, %1\;bfins\t%0, %2, 0, 16 -+ bfins\t%0, %2, 0, 16 -+ bfins\t%0, %1, 16, 16" -+ [(set_attr "length" "6, 4, 4") -+ (set_attr "type" "alu")]) -+ -+ -+;; Load the atomic operation description -+(include "sync.md") -+ -+;; Load the SIMD description -+(include "simd.md") -+ -+;; Load the FP coprocessor patterns -+(include "fpcp.md") -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-modes.def gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-modes.def ---- gcc-4.2.1/gcc/config/avr32/avr32-modes.def 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-modes.def 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1 @@ -+VECTOR_MODES (INT, 4); /* V4QI V2HI */ -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.opt gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.opt ---- gcc-4.2.1/gcc/config/avr32/avr32.opt 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.opt 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,73 @@ -+; Options for the ATMEL AVR32 port of the compiler. -+ -+; Copyright 2007 Atmel Corporation. -+; -+; This file is part of GCC. -+; -+; GCC is free software; you can redistribute it and/or modify it under -+; the terms of the GNU General Public License as published by the Free -+; Software Foundation; either version 2, or (at your option) any later -+; version. -+; -+; GCC is distributed in the hope that it will be useful, but WITHOUT ANY -+; WARRANTY; without even the implied warranty of MERCHANTABILITY or -+; FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License -+; for more details. -+; -+; You should have received a copy of the GNU General Public License -+; along with GCC; see the file COPYING. If not, write to the Free -+; Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA -+; 02110-1301, USA. -+ -+muse-rodata-section -+Target Report Mask(USE_RODATA_SECTION) -+Do not put readonly-data in .text section, but in .rodata. -+ -+mhard-float -+Target Report Undocumented Mask(HARD_FLOAT) -+Use floating point coprocessor instructions. -+ -+msoft-float -+Target Report Undocumented InverseMask(HARD_FLOAT, SOFT_FLOAT) -+Use software floating-point library for floating-point operations. -+ -+force-double-align -+Target Report RejectNegative Mask(FORCE_DOUBLE_ALIGN) -+Force double-word alignment for double-word memory accesses. -+ -+mno-init-got -+Target Report RejectNegative Mask(NO_INIT_GOT) -+Do not initialize GOT register before using it when compiling PIC code. -+ -+mrelax -+Target Report Mask(RELAX) -+Let invoked assembler and linker do relaxing (Enabled by default when optimization level is >1). -+ -+mmd-reorg-opt -+Target Report Undocumented Mask(MD_REORG_OPTIMIZATION) -+Perform machine dependent optimizations in reorg stage. -+ -+masm-addr-pseudos -+Target Report Mask(HAS_ASM_ADDR_PSEUDOS) -+Use assembler pseudo-instructions lda.w and call for handling direct addresses. (Enabled by default) -+ -+mpart= -+Target Report RejectNegative Joined Var(avr32_part_name) -+Specify the AVR32 part name -+ -+mcpu= -+Target Report RejectNegative Joined Undocumented Var(avr32_part_name) -+Specify the AVR32 part name (deprecated) -+ -+march= -+Target Report RejectNegative Joined Var(avr32_arch_name) -+Specify the AVR32 architecture name -+ -+mfast-float -+Target Report Mask(FAST_FLOAT) -+Enable fast floating-point library. Enabled by default if the -funsafe-math-optimizations switch is specified. -+ -+mimm-in-const-pool -+Target Report Var(avr32_imm_in_const_pool) Init(-1) -+Put large immediates in constant pool. This is enabled by default for archs with insn-cache. -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-protos.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-protos.h ---- gcc-4.2.1/gcc/config/avr32/avr32-protos.h 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-protos.h 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,185 @@ -+/* -+ Prototypes for exported functions defined in avr32.c -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ Initial porting by Anders �dland. -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+ -+#ifndef AVR32_PROTOS_H -+#define AVR32_PROTOS_H -+ -+extern const int swap_reg[]; -+ -+extern int avr32_valid_macmac_bypass (rtx, rtx); -+extern int avr32_valid_mulmac_bypass (rtx, rtx); -+ -+extern int avr32_decode_lcomm_symbol_offset (rtx, int *); -+extern void avr32_encode_lcomm_symbol_offset (tree, char *, int); -+ -+extern const char *avr32_strip_name_encoding (const char *); -+ -+extern rtx avr32_get_note_reg_equiv (rtx insn); -+ -+extern int avr32_use_return_insn (int iscond); -+ -+extern void avr32_make_reglist16 (int reglist16_vect, char *reglist16_string); -+ -+extern void avr32_make_reglist8 (int reglist8_vect, char *reglist8_string); -+extern void avr32_make_fp_reglist_w (int reglist_mask, char *reglist_string); -+extern void avr32_make_fp_reglist_d (int reglist_mask, char *reglist_string); -+ -+extern void avr32_output_return_instruction (int single_ret_inst, -+ int iscond, rtx cond, -+ rtx r12_imm); -+extern void avr32_expand_prologue (void); -+extern void avr32_set_return_address (rtx source, rtx scratch); -+ -+extern int avr32_hard_regno_mode_ok (int regno, enum machine_mode mode); -+extern int avr32_extra_constraint_s (rtx value, const int strict); -+extern int avr32_eh_return_data_regno (const int n); -+extern int avr32_initial_elimination_offset (const int from, const int to); -+extern rtx avr32_function_arg (CUMULATIVE_ARGS * cum, enum machine_mode mode, -+ tree type, int named); -+extern void avr32_init_cumulative_args (CUMULATIVE_ARGS * cum, tree fntype, -+ rtx libname, tree fndecl); -+extern void avr32_function_arg_advance (CUMULATIVE_ARGS * cum, -+ enum machine_mode mode, -+ tree type, int named); -+#ifdef ARGS_SIZE_RTX -+/* expr.h defines ARGS_SIZE_RTX and `enum direction'. */ -+extern enum direction avr32_function_arg_padding (enum machine_mode mode, -+ tree type); -+#endif /* ARGS_SIZE_RTX */ -+extern rtx avr32_function_value (tree valtype, tree func, bool outgoing); -+extern rtx avr32_libcall_value (enum machine_mode mode); -+extern int avr32_sched_use_dfa_pipeline_interface (void); -+extern bool avr32_return_in_memory (tree type, tree fntype); -+extern void avr32_regs_to_save (char *operand); -+extern void avr32_target_asm_function_prologue (FILE * file, -+ HOST_WIDE_INT size); -+extern void avr32_target_asm_function_epilogue (FILE * file, -+ HOST_WIDE_INT size); -+extern void avr32_trampoline_template (FILE * file); -+extern void avr32_initialize_trampoline (rtx addr, rtx fnaddr, -+ rtx static_chain); -+extern int avr32_legitimate_address (enum machine_mode mode, rtx x, -+ int strict); -+extern int avr32_legitimate_constant_p (rtx x); -+ -+extern int avr32_legitimate_pic_operand_p (rtx x); -+ -+extern rtx avr32_find_symbol (rtx x); -+extern void avr32_select_section (rtx exp, int reloc, int align); -+extern void avr32_encode_section_info (tree decl, rtx rtl, int first); -+extern void avr32_asm_file_end (FILE * stream); -+extern void avr32_asm_output_ascii (FILE * stream, char *ptr, int len); -+extern void avr32_asm_output_common (FILE * stream, const char *name, -+ int size, int rounded); -+extern void avr32_asm_output_label (FILE * stream, const char *name); -+extern void avr32_asm_declare_object_name (FILE * stream, char *name, -+ tree decl); -+extern void avr32_asm_globalize_label (FILE * stream, const char *name); -+extern void avr32_asm_weaken_label (FILE * stream, const char *name); -+extern void avr32_asm_output_external (FILE * stream, tree decl, -+ const char *name); -+extern void avr32_asm_output_external_libcall (FILE * stream, rtx symref); -+extern void avr32_asm_output_labelref (FILE * stream, const char *name); -+extern void avr32_notice_update_cc (rtx exp, rtx insn); -+extern void avr32_print_operand (FILE * stream, rtx x, int code); -+extern void avr32_print_operand_address (FILE * stream, rtx x); -+ -+extern int avr32_symbol (rtx x); -+ -+extern void avr32_select_rtx_section (enum machine_mode mode, rtx x, -+ unsigned HOST_WIDE_INT align); -+ -+extern int avr32_load_multiple_operation (rtx op, enum machine_mode mode); -+extern int avr32_store_multiple_operation (rtx op, enum machine_mode mode); -+ -+extern int avr32_const_ok_for_constraint_p (HOST_WIDE_INT value, char c, -+ const char *str); -+ -+extern bool avr32_cannot_force_const_mem (rtx x); -+ -+extern void avr32_init_builtins (void); -+ -+extern rtx avr32_expand_builtin (tree exp, rtx target, rtx subtarget, -+ enum machine_mode mode, int ignore); -+ -+extern bool avr32_must_pass_in_stack (enum machine_mode mode, tree type); -+ -+extern bool avr32_strict_argument_naming (CUMULATIVE_ARGS * ca); -+ -+extern bool avr32_pass_by_reference (CUMULATIVE_ARGS * cum, -+ enum machine_mode mode, -+ tree type, bool named); -+ -+extern rtx avr32_gen_load_multiple (rtx * regs, int count, rtx from, -+ int write_back, int in_struct_p, -+ int scalar_p); -+extern rtx avr32_gen_store_multiple (rtx * regs, int count, rtx to, -+ int in_struct_p, int scalar_p); -+extern int avr32_gen_movmemsi (rtx * operands); -+ -+extern int avr32_rnd_operands (rtx add, rtx shift); -+extern int avr32_adjust_insn_length (rtx insn, int length); -+ -+extern int symbol_mentioned_p (rtx x); -+extern int label_mentioned_p (rtx x); -+extern rtx legitimize_pic_address (rtx orig, enum machine_mode mode, rtx reg); -+extern int avr32_address_register_rtx_p (rtx x, int strict_p); -+extern int avr32_legitimate_index_p (enum machine_mode mode, rtx index, -+ int strict_p); -+ -+extern int avr32_const_double_immediate (rtx value); -+extern void avr32_init_expanders (void); -+extern rtx avr32_return_addr (int count, rtx frame); -+extern bool avr32_got_mentioned_p (rtx addr); -+ -+extern void avr32_final_prescan_insn (rtx insn, rtx * opvec, int noperands); -+ -+extern int avr32_expand_movcc (enum machine_mode mode, rtx operands[]); -+extern int avr32_expand_addcc (enum machine_mode mode, rtx operands[]); -+#ifdef RTX_CODE -+extern int avr32_expand_scc (RTX_CODE cond, rtx * operands); -+#endif -+ -+extern int avr32_store_bypass (rtx insn_out, rtx insn_in); -+extern int avr32_mul_waw_bypass (rtx insn_out, rtx insn_in); -+extern int avr32_valid_load_double_bypass (rtx insn_out, rtx insn_in); -+extern int avr32_valid_load_quad_bypass (rtx insn_out, rtx insn_in); -+extern rtx avr32_output_cmp (rtx cond, enum machine_mode mode, -+ rtx op0, rtx op1); -+ -+rtx get_next_insn_cond (rtx cur_insn); -+int set_next_insn_cond (rtx cur_insn, rtx cond); -+void avr32_override_options (void); -+void avr32_load_pic_register (void); -+void avr32_optimization_options (int level, int size); -+void avr32_split_const_expr (enum machine_mode mode, -+ enum machine_mode new_mode, -+ rtx expr, -+ rtx *split_expr); -+void avr32_get_intval (enum machine_mode mode, -+ rtx const_expr, -+ HOST_WIDE_INT *val); -+ -+ -+#endif /* AVR32_PROTOS_H */ -diff -Nrup gcc-4.2.1/gcc/config/avr32/crti.asm gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crti.asm ---- gcc-4.2.1/gcc/config/avr32/crti.asm 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crti.asm 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,64 @@ -+/* -+ Init/fini stuff for AVR32. -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+ -+/* The code in sections .init and .fini is supposed to be a single -+ regular function. The function in .init is called directly from -+ start in crt1.asm. The function in .fini is atexit()ed in crt1.asm -+ too. -+ -+ crti.asm contributes the prologue of a function to these sections, -+ and crtn.asm comes up the epilogue. STARTFILE_SPEC should list -+ crti.o before any other object files that might add code to .init -+ or .fini sections, and ENDFILE_SPEC should list crtn.o after any -+ such object files. */ -+ -+ .file "crti.asm" -+ -+ .section ".init" -+/* Just load the GOT */ -+ .align 2 -+ .global _init -+_init: -+ stm --sp, r6, lr -+ lddpc r6, 1f -+0: -+ rsub r6, pc -+ rjmp 2f -+ .align 2 -+1: .long 0b - _GLOBAL_OFFSET_TABLE_ -+2: -+ -+ .section ".fini" -+/* Just load the GOT */ -+ .align 2 -+ .global _fini -+_fini: -+ stm --sp, r6, lr -+ lddpc r6, 1f -+0: -+ rsub r6, pc -+ rjmp 2f -+ .align 2 -+1: .long 0b - _GLOBAL_OFFSET_TABLE_ -+2: -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/crtn.asm gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crtn.asm ---- gcc-4.2.1/gcc/config/avr32/crtn.asm 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crtn.asm 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,44 @@ -+/* Copyright (C) 2001 Free Software Foundation, Inc. -+ Written By Nick Clifton -+ -+ This file is free software; you can redistribute it and/or modify it -+ under the terms of the GNU General Public License as published by the -+ Free Software Foundation; either version 2, or (at your option) any -+ later version. -+ -+ In addition to the permissions in the GNU General Public License, the -+ Free Software Foundation gives you unlimited permission to link the -+ compiled version of this file with other programs, and to distribute -+ those programs without any restriction coming from the use of this -+ file. (The General Public License restrictions do apply in other -+ respects; for example, they cover modification of the file, and -+ distribution when not linked into another program.) -+ -+ This file is distributed in the hope that it will be useful, but -+ WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -+ General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; see the file COPYING. If not, write to -+ the Free Software Foundation, 59 Temple Place - Suite 330, -+ Boston, MA 02111-1307, USA. -+ -+ As a special exception, if you link this library with files -+ compiled with GCC to produce an executable, this does not cause -+ the resulting executable to be covered by the GNU General Public License. -+ This exception does not however invalidate any other reasons why -+ the executable file might be covered by the GNU General Public License. -+*/ -+ -+ -+ -+ -+ .file "crtn.asm" -+ -+ .section ".init" -+ ldm sp++, r6, pc -+ -+ .section ".fini" -+ ldm sp++, r6, pc -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/fpcp.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/fpcp.md ---- gcc-4.2.1/gcc/config/avr32/fpcp.md 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/fpcp.md 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,551 @@ -+;; AVR32 machine description file for Floating-Point instructions. -+;; Copyright 2003-2006 Atmel Corporation. -+;; -+;; Written by Ronny Pedersen, Atmel Norway, -+;; -+;; This file is part of GCC. -+;; -+;; This program is free software; you can redistribute it and/or modify -+;; it under the terms of the GNU General Public License as published by -+;; the Free Software Foundation; either version 2 of the License, or -+;; (at your option) any later version. -+;; -+;; This program is distributed in the hope that it will be useful, -+;; but WITHOUT ANY WARRANTY; without even the implied warranty of -+;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+;; GNU General Public License for more details. -+;; -+;; You should have received a copy of the GNU General Public License -+;; along with this program; if not, write to the Free Software -+;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. -+ -+;; -*- Mode: Scheme -*- -+ -+;;****************************************************************************** -+;; Automaton pipeline description for floating-point coprocessor insns -+;;****************************************************************************** -+(define_cpu_unit "fid,fm1,fm2,fm3,fm4,fwb,fcmp,fcast" "avr32_ap") -+ -+(define_insn_reservation "fmv_op" 1 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fmv")) -+ "is,da,d,fid,fwb") -+ -+(define_insn_reservation "fmul_op" 5 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fmul")) -+ "is,da,d,fid,fm1,fm2,fm3,fm4,fwb") -+ -+(define_insn_reservation "fcmps_op" 1 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fcmps")) -+ "is,da,d,fid,fcmp") -+ -+(define_insn_reservation "fcmpd_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fcmpd")) -+ "is,da,d,fid*2,fcmp") -+ -+(define_insn_reservation "fcast_op" 3 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fcast")) -+ "is,da,d,fid,fcmp,fcast,fwb") -+ -+(define_insn_reservation "fmvcpu_op" 2 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fmvcpu")) -+ "is,da,d") -+ -+(define_insn_reservation "fldd_op" 1 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fldd")) -+ "is,da,d,fwb") -+ -+(define_insn_reservation "flds_op" 1 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "flds")) -+ "is,da,d,fwb") -+ -+(define_insn_reservation "fsts_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fsts")) -+ "is,da*2,d") -+ -+(define_insn_reservation "fstd_op" 0 -+ (and (eq_attr "pipeline" "ap") -+ (eq_attr "type" "fstd")) -+ "is,da*2,d") -+ -+ -+(define_insn "*movsf_fpcp" -+ [(set (match_operand:SF 0 "nonimmediate_operand" "=f,f,r,f,m,r,r,r,m") -+ (match_operand:SF 1 "general_operand" " f,r,f,m,f,r,G,m,r"))] -+ "TARGET_HARD_FLOAT" -+ "@ -+ fmov.s\t%0, %1 -+ fmov.s\t%0, %1 -+ fmov.s\t%0, %1 -+ fld.s\t%0, %1 -+ fst.s\t%0, %1 -+ mov\t%0, %1 -+ mov\t%0, %1 -+ ld.w\t%0, %1 -+ st.w\t%0, %1" -+ [(set_attr "length" "4,4,4,4,4,2,4,4,4") -+ (set_attr "type" "fmv,flds,fmvcpu,flds,fsts,alu,alu,load,store")]) -+ -+(define_insn_and_split "*movdf_fpcp" -+ [(set (match_operand:DF 0 "nonimmediate_operand" "=f,f,r,f,m,r,r,m") -+ (match_operand:DF 1 "general_operand" " f,r,f,m,f,r,m,r"))] -+ "TARGET_HARD_FLOAT" -+ "@ -+ fmov.d\t%0, %1 -+ fmov.d\t%0, %1 -+ fmov.d\t%0, %1 -+ fld.d\t%0, %1 -+ fst.d\t%0, %1 -+ mov\t%0, %1\;mov\t%m0, %m1 -+ ld.d\t%0, %1 -+ st.d\t%0, %1" -+ -+ "TARGET_HARD_FLOAT -+ && reload_completed -+ && (REG_P(operands[0]) && (REGNO_REG_CLASS(REGNO(operands[0])) == GENERAL_REGS)) -+ && (REG_P(operands[1]) && (REGNO_REG_CLASS(REGNO(operands[1])) == GENERAL_REGS))" -+ [(set (match_dup 0) (match_dup 1)) -+ (set (match_dup 2) (match_dup 3))] -+ " -+ { -+ operands[2] = gen_highpart (SImode, operands[0]); -+ operands[0] = gen_lowpart (SImode, operands[0]); -+ operands[3] = gen_highpart(SImode, operands[1]); -+ operands[1] = gen_lowpart(SImode, operands[1]); -+ } -+ " -+ -+ [(set_attr "length" "4,4,4,4,4,4,4,4") -+ (set_attr "type" "fmv,fldd,fmvcpu,fldd,fstd,alu2,load2,store2")]) -+ -+ -+(define_insn "mulsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fmul.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "nmulsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f"))))] -+ "TARGET_HARD_FLOAT" -+ "fnmul.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") -+ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") -+ (match_operand:SF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:SF 3 "avr32_fp_register_operand" "") -+ (neg:SF (match_dup 0)))] -+ "TARGET_HARD_FLOAT && -+ (peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[3]) == REGNO(operands[0])))" -+ [(set (match_dup 3) -+ (neg:SF (mult:SF (match_dup 1) -+ (match_dup 2))))] -+) -+ -+ -+(define_insn "macsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (plus:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f")) -+ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fmac.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "nmacsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (plus:SF (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f"))) -+ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fnmac.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") -+ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") -+ (match_operand:SF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:SF 3 "avr32_fp_register_operand" "") -+ (minus:SF -+ (match_dup 3) -+ (match_dup 0)))] -+ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 3) -+ (plus:SF (neg:SF (mult:SF (match_dup 1) -+ (match_dup 2))) -+ (match_dup 3)))] -+) -+ -+ -+(define_insn "msubacsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (minus:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f")) -+ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fmsc.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") -+ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") -+ (match_operand:SF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:SF 3 "avr32_fp_register_operand" "") -+ (minus:SF -+ (match_dup 0) -+ (match_dup 3)))] -+ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 3) -+ (minus:SF (mult:SF (match_dup 1) -+ (match_dup 2)) -+ (match_dup 3)))] -+) -+ -+(define_insn "nmsubacsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (minus:SF (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f"))) -+ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fnmsc.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+ -+ -+(define_insn "addsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (plus:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fadd.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "subsf3" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (minus:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") -+ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fsub.s\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+ -+(define_insn "negsf2" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (neg:SF (match_operand:SF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fneg.s\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmv")]) -+ -+(define_insn "abssf2" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (abs:SF (match_operand:SF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fabs.s\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmv")]) -+ -+(define_insn "truncdfsf2" -+ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") -+ (float_truncate:SF -+ (match_operand:DF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fcastd.s\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fcast")]) -+ -+(define_insn "extendsfdf2" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (float_extend:DF -+ (match_operand:SF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fcasts.d\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fcast")]) -+ -+(define_insn "muldf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fmul.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "nmuldf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f"))))] -+ "TARGET_HARD_FLOAT" -+ "fnmul.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") -+ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") -+ (match_operand:DF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:DF 3 "avr32_fp_register_operand" "") -+ (neg:DF (match_dup 0)))] -+ "TARGET_HARD_FLOAT && -+ (peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[3]) == REGNO(operands[0])))" -+ [(set (match_dup 3) -+ (neg:DF (mult:DF (match_dup 1) -+ (match_dup 2))))] -+) -+ -+(define_insn "macdf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (plus:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f")) -+ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fmac.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "msubacdf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (minus:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f")) -+ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fmsc.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") -+ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") -+ (match_operand:DF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:DF 3 "avr32_fp_register_operand" "") -+ (minus:DF -+ (match_dup 0) -+ (match_dup 3)))] -+ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 3) -+ (minus:DF (mult:DF (match_dup 1) -+ (match_dup 2)) -+ (match_dup 3)))] -+ ) -+ -+(define_insn "nmsubacdf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (minus:DF (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f"))) -+ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fnmsc.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "nmacdf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (plus:DF (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f"))) -+ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] -+ "TARGET_HARD_FLOAT" -+ "fnmac.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_peephole2 -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") -+ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") -+ (match_operand:DF 2 "avr32_fp_register_operand" ""))) -+ (set (match_operand:DF 3 "avr32_fp_register_operand" "") -+ (minus:DF -+ (match_dup 3) -+ (match_dup 0)))] -+ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" -+ [(set (match_dup 3) -+ (plus:DF (neg:DF (mult:DF (match_dup 1) -+ (match_dup 2))) -+ (match_dup 3)))] -+) -+ -+(define_insn "adddf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (plus:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fadd.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "subdf3" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (minus:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") -+ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fsub.d\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmul")]) -+ -+(define_insn "negdf2" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (neg:DF (match_operand:DF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fneg.d\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmv")]) -+ -+(define_insn "absdf2" -+ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") -+ (abs:DF (match_operand:DF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ "fabs.d\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmv")]) -+ -+ -+(define_expand "cmpdf" -+ [(set (cc0) -+ (compare:DF -+ (match_operand:DF 0 "general_operand" "") -+ (match_operand:DF 1 "general_operand" "")))] -+ "TARGET_HARD_FLOAT" -+ "{ -+ rtx tmpreg; -+ if ( !REG_P(operands[0]) ) -+ operands[0] = force_reg(DFmode, operands[0]); -+ -+ if ( !REG_P(operands[1]) ) -+ operands[1] = force_reg(DFmode, operands[1]); -+ -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ -+ emit_insn(gen_cmpdf_internal(operands[0], operands[1])); -+ -+ tmpreg = gen_reg_rtx(SImode); -+ emit_insn(gen_fpcc_to_reg(tmpreg)); -+ emit_insn(gen_reg_to_cc(tmpreg)); -+ -+ DONE; -+ }" -+) -+ -+(define_insn "cmpdf_internal" -+ [(set (reg:CC FPCC_REGNUM) -+ (compare:CC -+ (match_operand:DF 0 "avr32_fp_register_operand" "f") -+ (match_operand:DF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ { -+ if (!rtx_equal_p(cc_prev_status.mdep.fpvalue, SET_SRC(PATTERN (insn))) ) -+ return "fcmp.d\t%0, %1"; -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "type" "fcmpd") -+ (set_attr "cc" "fpcompare")]) -+ -+(define_expand "cmpsf" -+ [(set (cc0) -+ (compare:SF -+ (match_operand:SF 0 "general_operand" "") -+ (match_operand:SF 1 "general_operand" "")))] -+ "TARGET_HARD_FLOAT" -+ "{ -+ rtx tmpreg; -+ if ( !REG_P(operands[0]) ) -+ operands[0] = force_reg(SFmode, operands[0]); -+ -+ if ( !REG_P(operands[1]) ) -+ operands[1] = force_reg(SFmode, operands[1]); -+ -+ avr32_compare_op0 = operands[0]; -+ avr32_compare_op1 = operands[1]; -+ -+ emit_insn(gen_cmpsf_internal(operands[0], operands[1])); -+ -+ tmpreg = gen_reg_rtx(SImode); -+ emit_insn(gen_fpcc_to_reg(tmpreg)); -+ emit_insn(gen_reg_to_cc(tmpreg)); -+ -+ DONE; -+ }" -+) -+ -+(define_insn "cmpsf_internal" -+ [(set (reg:CC FPCC_REGNUM) -+ (compare:CC -+ (match_operand:SF 0 "avr32_fp_register_operand" "f") -+ (match_operand:SF 1 "avr32_fp_register_operand" "f")))] -+ "TARGET_HARD_FLOAT" -+ { -+ if (!rtx_equal_p(cc_prev_status.mdep.fpvalue, SET_SRC(PATTERN (insn))) ) -+ return "fcmp.s\t%0, %1"; -+ return ""; -+ } -+ [(set_attr "length" "4") -+ (set_attr "type" "fcmps") -+ (set_attr "cc" "fpcompare")]) -+ -+(define_insn "fpcc_to_reg" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (unspec:SI [(reg:CC FPCC_REGNUM)] -+ UNSPEC_FPCC_TO_REG))] -+ "TARGET_HARD_FLOAT" -+ "fmov.s\t%0, fsr" -+ [(set_attr "length" "4") -+ (set_attr "type" "fmvcpu")]) -+ -+(define_insn "reg_to_cc" -+ [(set (cc0) -+ (unspec:SI [(match_operand:SI 0 "register_operand" "r")] -+ UNSPEC_REG_TO_CC))] -+ "TARGET_HARD_FLOAT" -+ "musfr\t%0" -+ [(set_attr "length" "2") -+ (set_attr "type" "alu") -+ (set_attr "cc" "from_fpcc")]) -+ -+(define_insn "stm_fp" -+ [(unspec [(match_operand 0 "register_operand" "r") -+ (match_operand 1 "const_int_operand" "") -+ (match_operand 2 "const_int_operand" "")] -+ UNSPEC_STMFP)] -+ "TARGET_HARD_FLOAT" -+ { -+ int cop_reglist = INTVAL(operands[1]); -+ -+ if (INTVAL(operands[2]) != 0) -+ return "stcm.w\tcp0, --%0, %C1"; -+ else -+ return "stcm.w\tcp0, %0, %C1"; -+ -+ if ( cop_reglist & ~0xff ){ -+ operands[1] = GEN_INT(cop_reglist & ~0xff); -+ if (INTVAL(operands[2]) != 0) -+ return "stcm.d\tcp0, --%0, %D1"; -+ else -+ return "stcm.d\tcp0, %0, %D1"; -+ } -+ } -+ [(set_attr "type" "fstm") -+ (set_attr "length" "4") -+ (set_attr "cc" "none")]) -diff -Nrup gcc-4.2.1/gcc/config/avr32/lib1funcs.S gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib1funcs.S ---- gcc-4.2.1/gcc/config/avr32/lib1funcs.S 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib1funcs.S 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,2589 @@ -+ -+/* Adjust the unpacked double number if it is a subnormal number. -+ The exponent and mantissa pair are stored -+ in [mant_hi,mant_lo] and [exp]. A register with the correct sign bit in -+ the MSB is passed in [sign]. Needs two scratch -+ registers [scratch1] and [scratch2]. An adjusted and packed double float -+ is present in [mant_hi,mant_lo] after macro has executed */ -+.macro adjust_subnormal_df exp, mant_lo, mant_hi, sign, scratch1, scratch2 -+ /* We have an exponent which is <=0 indicating a subnormal number -+ As it should be stored as if the exponent was 1 (although the -+ exponent field is all zeros to indicate a subnormal number) -+ we have to shift down the mantissa to its correct position. */ -+ neg \exp -+ sub \exp,-1 /* amount to shift down */ -+ cp.w \exp,54 -+ brlo 50f /* if more than 53 shift steps, the -+ entire mantissa will disappear -+ without any rounding to occur */ -+ mov \mant_hi, 0 -+ mov \mant_lo, 0 -+ rjmp 52f -+50: -+ sub \exp,-10 /* do the shift to position the -+ mantissa at the same time -+ note! this does not include the -+ final 1 step shift to add the sign */ -+ -+ /* when shifting, save all shifted out bits in [scratch2]. we may need to -+ look at them to make correct rounding. */ -+ -+ rsub \scratch1,\exp,32 /* get inverted shift count */ -+ cp.w \exp,32 /* handle shifts >= 32 separately */ -+ brhs 51f -+ -+ /* small (<32) shift amount, both words are part of the shift */ -+ lsl \scratch2,\mant_lo,\scratch1 /* save bits to shift out from lsw*/ -+ lsl \scratch1,\mant_hi,\scratch1 /* get bits from msw destined for lsw*/ -+ lsr \mant_lo,\mant_lo,\exp /* shift down lsw */ -+ lsr \mant_hi,\mant_hi,\exp /* shift down msw */ -+ or \mant_hi,\scratch1 /* add bits from msw with prepared lsw */ -+ rjmp 50f -+ -+ /* large (>=32) shift amount, only lsw will have bits left after shift. -+ note that shift operations will use ((shift count) mod 32) so -+ we do not need to subtract 32 from shift count. */ -+51: -+ lsl \scratch2,\mant_hi,\scratch1 /* save bits to shift out from msw */ -+ or \scratch2,\mant_lo /* also save all bits from lsw */ -+ mov \mant_lo,\mant_hi /* msw -> lsw (i.e. "shift 32 first") */ -+ mov \mant_hi,0 /* clear msw */ -+ lsr \mant_lo,\mant_lo,\exp /* make rest of shift inside lsw */ -+ -+50: -+ /* result is almost ready to return, except that least significant bit -+ and the part we already shifted out may cause the result to be -+ rounded */ -+ bld \mant_lo,0 /* get bit to be shifted out */ -+ brcc 51f /* if bit was 0, no rounding */ -+ -+ /* msb of part to remove is 1, so rounding depends on rest of bits */ -+ tst \scratch2,\scratch2 /* get shifted out tail */ -+ brne 50f /* if rest > 0, do round */ -+ bld \mant_lo,1 /* we have to look at lsb in result */ -+ brcc 51f /* if lsb is 0, don't round */ -+ -+50: -+ /* subnormal result requires rounding -+ rounding may cause subnormal to become smallest normal number -+ luckily, smallest normal number has exactly the representation -+ we got by rippling a one bit up from mantissa into exponent field. */ -+ sub \mant_lo,-1 -+ subcc \mant_hi,-1 -+ -+51: -+ /* shift and return packed double with correct sign */ -+ rol \sign -+ ror \mant_hi -+ ror \mant_lo -+52: -+.endm -+ -+ -+/* Adjust subnormal single float number with exponent [exp] -+ and mantissa [mant] and round. */ -+.macro adjust_subnormal_sf sf, exp, mant, sign, scratch -+ /* subnormal number */ -+ rsub \exp,\exp, 1 /* shift amount */ -+ cp.w \exp, 25 -+ movhs \mant, 0 -+ brhs 90f /* Return zero */ -+ rsub \scratch, \exp, 32 -+ lsl \scratch, \mant,\scratch/* Check if there are any bits set -+ in the bits discarded in the mantissa */ -+ srne \scratch /* If so set the lsb of the shifted mantissa */ -+ lsr \mant,\mant,\exp /* Shift the mantissa */ -+ or \mant, \scratch /* Round lsb if any bits were shifted out */ -+ /* Rounding : For explaination, see round_sf. */ -+ mov \scratch, 0x7f /* Set rounding constant */ -+ bld \mant, 8 -+ subeq \scratch, -1 /* For odd numbers use rounding constant 0x80 */ -+ add \mant, \scratch /* Add rounding constant to mantissa */ -+ /* We can't overflow because mantissa is at least shifted one position -+ to the right so the implicit bit is zero. We can however get the implicit -+ bit set after rounding which means that we have the lowest normal number -+ but this is ok since this bit has the same position as the LSB of the -+ exponent */ -+ lsr \sf, \mant, 7 -+ /* Rotate in sign */ -+ lsl \sign, 1 -+ ror \sf -+90: -+.endm -+ -+ -+/* Round the unpacked df number with exponent [exp] and -+ mantissa [mant_hi, mant_lo]. Uses scratch register -+ [scratch] */ -+.macro round_df exp, mant_lo, mant_hi, scratch -+ mov \scratch, 0x3ff /* Rounding constant */ -+ bld \mant_lo,11 /* Check if lsb in the final result is -+ set */ -+ subeq \scratch, -1 /* Adjust rounding constant to 0x400 -+ if rounding 0.5 upwards */ -+ add \mant_lo, \scratch /* Round */ -+ acr \mant_hi /* If overflowing we know that -+ we have all zeros in the bits not -+ scaled out so we can leave them -+ but we must increase the exponent with -+ two since we had an implicit bit -+ which is lost + the extra overflow bit */ -+ subcs \exp, -2 /* Update exponent */ -+.endm -+ -+/* Round single float number stored in [mant] and [exp] */ -+.macro round_sf exp, mant, scratch -+ /* Round: -+ For 0.5 we round to nearest even integer -+ for all other cases we round to nearest integer. -+ This means that if the digit left of the "point" (.) -+ is 1 we can add 0x80 to the mantissa since the -+ corner case 0x180 will round up to 0x200. If the -+ digit left of the "point" is 0 we will have to -+ add 0x7f since this will give 0xff and hence a -+ truncation/rounding downwards for the corner -+ case when the 9 lowest bits are 0x080 */ -+ mov \scratch, 0x7f /* Set rounding constant */ -+ /* Check if the mantissa is even or odd */ -+ bld \mant, 8 -+ subeq \scratch, -1 /* Rounding constant should be 0x80 */ -+ add \mant, \scratch -+ subcs \exp, -2 /* Adjust exponent if we overflowed */ -+.endm -+ -+/* Scale mantissa [mant_hi, mant_lo] with amount [shift_count]. -+ Uses scratch registers [scratch1] and [scratch2] */ -+.macro scale_df shift_count, mant_lo, mant_hi, scratch1, scratch2 -+ /* Scale [mant_hi, mant_lo] with shift_amount. -+ Must not forget the sticky bits we intend to shift out. */ -+ -+ rsub \scratch1,\shift_count,32/* get (32 - shift count) -+ (if shift count > 32 we get a -+ negative value, but that will -+ work as well in the code below.) */ -+ -+ cp.w \shift_count,32 /* handle shifts >= 32 separately */ -+ brhs 70f -+ -+ /* small (<32) shift amount, both words are part of the shift -+ first remember whether part that is lost contains any 1 bits ... */ -+ lsl \scratch2,\mant_lo,\scratch1 /*shift away bits that are part of -+ final mantissa. only part that goes -+ to scratch2 are bits that will be lost */ -+ -+ /* ... and now to the actual shift */ -+ lsl \scratch1,\mant_hi,\scratch1 /* get bits from msw destined for lsw*/ -+ lsr \mant_lo,\mant_lo,\shift_count /* shift down lsw of mantissa */ -+ lsr \mant_hi,\mant_hi,\shift_count /* shift down msw of mantissa */ -+ or \mant_lo,\scratch1 /* combine these bits with prepared lsw*/ -+ rjmp 71f -+ -+ /* large (>=32) shift amount, only lsw will have bits left after shift. -+ note that shift operations will use ((shift count) mod 32) so -+ we do not need to subtract 32 from shift count. */ -+70: -+ /* first remember whether part that is lost contains any 1 bits ... */ -+ lsl \scratch2,\mant_hi,\scratch1 /* save all lost bits from msw */ -+ or \scratch2,\mant_lo /* also save lost bits (all) from lsw -+ now scratch2<>0 if we lose any bits */ -+ -+ /* ... and now to the actual shift */ -+ mov \mant_lo,\mant_hi /* msw -> lsw (i.e. "shift 32 first")*/ -+ mov \mant_hi,0 /* clear msw */ -+ lsr \mant_lo,\mant_lo,\shift_count /* make rest of shift inside lsw*/ -+ -+71: -+ cp.w \scratch2,0 /* if any '1' bit in part we lost ...*/ -+ breq 70f -+ -+ sbr \mant_lo,0 /* ... we need to set sticky bit*/ -+70: -+.endm -+ -+/* Unpack exponent and mantissa from the double number -+ stored in [df_hi,df_lo]. The exponent is stored in [exp] -+ while the mantissa is stored in [df_hi,df_lo]. */ -+ -+.macro unpack_df exp, df_lo, df_hi -+ lsr \exp, \df_hi,21 /* Extract exponent */ -+ lsl \df_hi,10 /* Get mantissa */ -+ or \df_hi,\df_hi,\df_lo>>21 -+ lsl \df_lo,11 -+ -+ neg \exp /* Fix implicit bit */ -+ bst \df_hi,31 -+ subeq \exp,1 -+ neg \exp /* negate back exponent */ -+ .endm -+ -+/* Unpack exponent and mantissa from the single float number -+ stored in [sf]. The exponent is stored in [exp] -+ while the mantissa is stored in [sf]. */ -+.macro unpack_sf exp, sf -+ lsr \exp, \sf, 24 -+ brne 80f -+ /* Fix subnormal number */ -+ lsl \sf,7 -+ clz \exp,\sf -+ lsl \sf,\sf,\exp -+ rsub \exp,\exp,1 -+ rjmp 81f -+80: -+ lsl \sf,7 -+ sbr \sf, 31 /*Implicit bit*/ -+81: -+.endm -+ -+ -+ -+/* Pack a single float number stored in [mant] and [exp] -+ into a single float number in [sf] */ -+.macro pack_sf sf, exp, mant -+ bld \mant,31 /* implicit bit to z */ -+ subne \exp,1 /* if subnormal (implicit bit 0) -+ adjust exponent to storage format */ -+ -+ lsr \sf, \mant, 7 -+ bfins \sf, \exp, 24, 8 -+.endm -+ -+/* Pack exponent [exp] and mantissa [mant_hi, mant_lo] -+ into [df_hi, df_lo]. [df_hi] is shifted -+ one bit up so the sign bit can be shifted into it */ -+ -+.macro pack_df exp, mant_lo, mant_hi, df_lo, df_hi -+ bld \mant_hi,31 /* implicit bit to z */ -+ subne \exp,1 /* if subnormal (implicit bit 0) -+ adjust exponent to storage format */ -+ -+ lsr \mant_lo,11 /* shift back lsw */ -+ or \df_lo,\mant_lo,\mant_hi<<21 /* combine with low bits from msw */ -+ lsl \mant_hi,1 /* get rid of implicit bit */ -+ lsr \mant_hi,11 /* shift back msw except for one step*/ -+ or \df_hi,\mant_hi,\exp<<21 /* combine msw with exponent */ -+.endm -+ -+/* Normalize single float number stored in [mant] and [exp] -+ using scratch register [scratch] */ -+.macro normalize_sf exp, mant, scratch -+ /* Adjust exponent and mantissa */ -+ clz \scratch, \mant -+ sub \exp, \scratch -+ lsl \mant, \mant, \scratch -+.endm -+ -+/* Normalize the exponent and mantissa pair stored -+ in [mant_hi,mant_lo] and [exp]. Needs two scratch -+ registers [scratch1] and [scratch2]. */ -+.macro normalize_df exp, mant_lo, mant_hi, scratch1, scratch2 -+ clz \scratch1,\mant_hi /* Check if we have zeros in high bits */ -+ breq 80f /* No need for scaling if no zeros in high bits */ -+ brcs 81f /* Check for all zeros */ -+ -+ /* shift amount is smaller than 32, and involves both msw and lsw*/ -+ rsub \scratch2,\scratch1,32 /* shift mantissa */ -+ lsl \mant_hi,\mant_hi,\scratch1 -+ lsr \scratch2,\mant_lo,\scratch2 -+ or \mant_hi,\scratch2 -+ lsl \mant_lo,\mant_lo,\scratch1 -+ sub \exp,\scratch1 /* adjust exponent */ -+ rjmp 80f /* Finished */ -+81: -+ /* shift amount is greater than 32 */ -+ clz \scratch1,\mant_lo /* shift mantissa */ -+ movcs \scratch1, 0 -+ subcc \scratch1,-32 -+ mov \mant_hi,\mant_lo -+ lsl \mant_hi,\mant_hi,\scratch1 -+ mov \mant_lo,0 -+ sub \exp,\scratch1 /* adjust exponent */ -+80: -+.endm -+ -+ -+/* Fast but approximate multiply of two 64-bit numbers to give a 64 bit result. -+ The multiplication of [al]x[bl] is discarded. -+ Operands in [ah], [al], [bh], [bl]. -+ Scratch registers in [sh], [sl]. -+ Returns results in registers [rh], [rl].*/ -+.macro mul_approx_df ah, al, bh, bl, rh, rl, sh, sl -+ mulu.d \sl, \ah, \bl -+ macu.d \sl, \al, \bh -+ mulu.d \rl, \ah, \bh -+ add \rl, \sh -+ acr \rh -+.endm -+ -+ -+ -+ -+#ifdef L_avr32_f64_mul -+ .align 2 -+ .global __avr32_f64_mul -+ .type __avr32_f64_mul,@function -+ -+ -+__avr32_f64_mul: -+ or r12, r10, r11 << 1 -+ breq __avr32_f64_mul_op1_zero -+ -+ stm --sp, r5,r6,r7,lr -+ /* op1 in {r11,r10}*/ -+ /* op2 in {r9,r8}*/ -+ eor lr, r11, r9 /* MSB(lr) = Sign(op1) ^ Sign(op2) */ -+ -+ /* Unpack op1 */ -+ /* exp: r7 */ -+ /* sf: r11, r10 */ -+ lsr r7, r11, 20 /* Extract exponent */ -+ -+ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r11, r11, r10>>21 -+ lsl r10, 11 -+ sbr r11, 31 /* Insert implicit bit */ -+ -+ cbr r7, 11 /* Clear sign bit */ -+ /* Check if normalization is needed */ -+ breq __avr32_f64_mul_op1_subnormal /*If number is subnormal, normalize it */ -+ -+22: -+ /* Unpack op2 */ -+ /* exp: r6 */ -+ /* sf: r9, r8 */ -+ lsr r6, r9, 20 /* Extract exponent */ -+ -+ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>21 -+ lsl r8, 11 -+ sbr r9, 31 /* Insert implicit bit */ -+ -+ cbr r6, 11 /* Clear sign bit */ -+ /* Check if normalization is needed */ -+ breq __avr32_f64_mul_op2_subnormal /*If number is subnormal, normalize it */ -+23: -+ -+ /* Check if any operands are NaN or INF */ -+ cp r7, 0x7ff -+ breq __avr32_f64_mul_op_nan_or_inf /* Check op1 for NaN or Inf */ -+ cp r6, 0x7ff -+ breq __avr32_f64_mul_op_nan_or_inf /* Check op2 for NaN or Inf */ -+ -+ -+ /* Calculate new exponent in r12*/ -+ add r12, r7, r6 -+ sub r12, (1023-1) -+ -+ /* Do the multiplication using approximate calculation. -+ Place result in r11, r10. Use r7, r6 as scratch registers */ -+ mul_approx_df r11 /*ah*/, r10 /*al*/, r9 /*bh*/, r8 /*bl*/, r11 /*rh*/, r10 /*rl*/, r7 /*sh*/, r6 /*sl*/ -+ -+ /* Check if result is zero */ -+ breq __avr32_f64_mul_res_subnormal -+ -+ /* Adjust exponent and mantissa */ -+ /* [r12]:exp, [r11, r10]:mant, [r9,r8,r7]:scratch*/ -+ /* Mantissa may be of the format 0.xxxx or 1.xxxx. */ -+ /* In the first case, shift one pos to left.*/ -+ sub r9, r12, 1 -+ mov r8, r11 -+ lsl r7, r10, 1 -+ rol r8 -+ bld r11, 31 -+ movne r12, r9 -+ movne r11, r8 -+ movne r10, r7 -+ cp r12, 0 -+ breq __avr32_f64_mul_res_subnormal /*Result was subnormal. Flush-to-zero and return zero*/ -+ -+ /* Check for Inf. */ -+ cp.w r12, 0x7ff -+ brge __avr32_f64_mul_res_inf -+ -+ /* Result was not subnormal. Perform rounding. */ -+ /* Because of performance optimization, we have no sticky bit, */ -+ /* so round-to-even won't work as specified in the IEEE standard.*/ -+ /* [r12]:exp, [r11, r10]:mant */ -+ /* Mantissa is in 0.64 format. Round by adding 1<<(64-(52+2))=1<<10*/ -+ /* That is, 1 in the MSB of the part that will be discarded in final packing.*/ -+ mov r9, (1<<10) -+ add r10, r9 -+ acr r11 -+ /* Adjust exponent if we overflowed.*/ -+ subcs r12, -1 -+ -+ -+ /* Pack final result*/ -+ /* Input: [r12]:exp, [r11, r10]:mant */ -+ /* Result in [r11,r10] */ -+ /* Insert mantissa */ -+ cbr r11, 31 /*Clear implicit bit*/ -+ lsr r10, 11 -+ or r10, r10, r11<<21 -+ lsr r11, 11 -+ /* Insert exponent and sign bit*/ -+ or r11, r11, r12<<20 -+ bld lr, 31 -+ bst r11, 31 -+ -+ /* Return result in [r11,r10] */ -+ ldm sp++, r5, r6, r7,pc -+ -+ -+__avr32_f64_mul_op1_subnormal: -+ cbr r11, 31 /* Clear implicit bit. */ -+ normalize_df r7 /*exp*/, r10, r11 /*Mantissa*/, r5, r12 /*scratch*/ -+ rjmp 22b -+ -+__avr32_f64_mul_op2_subnormal: -+ cbr r9, 31 /* Clear implicit bit. */ -+ normalize_df r6 /*exp*/, r8, r9 /*Mantissa*/, r5, r12 /*scratch*/ -+ rjmp 23b -+ -+ -+__avr32_f64_mul_op_nan_or_inf: -+ /* Same code for OP1 and OP2*/ -+ /* Since we are here, at least one of the OPs were NaN or INF*/ -+ /* Shift out implicit bit of both operands' mantissa */ -+ lsl r11, 1 -+ lsl r9, 1 -+ /* Merge the regs in each operand to check for zero*/ -+ or r11, r10 /* op1 */ -+ or r9, r8 /* op2 */ -+ /* Check if op1 is NaN or INF */ -+ cp r7, 0x7ff -+ brne __avr32_f64_mul_op1_not_naninf -+ /* op1 was NaN or INF.*/ -+ cp r11, 0 -+ brne __avr32_f64_mul_res_nan /* op1 was NaN. Result will be NaN*/ -+ /*op1 was INF. check if op2 is NaN or INF*/ -+ cp r6, 0x7ff -+ brne __avr32_f64_mul_res_inf /*op1 was INF, op2 was neither NaN nor INF*/ -+ /* op1 is INF, op2 is either NaN or INF*/ -+ cp r9, 0 -+ breq __avr32_f64_mul_res_inf /*op2 was also INF*/ -+ rjmp __avr32_f64_mul_res_nan /*op2 was NaN*/ -+ -+__avr32_f64_mul_op1_not_naninf: -+ /* op1 was not NaN nor INF. Then op2 must be NaN or INF*/ -+ cp r9, 0 -+ breq __avr32_f64_mul_res_inf /*op2 was INF, return INF*/ -+ rjmp __avr32_f64_mul_res_nan /*else return NaN*/ -+ -+__avr32_f64_mul_res_subnormal:/* Multiply result was subnormal. Return zero. */ -+ mov r11, lr /*Get correct sign*/ -+ andh r11, 0x8000, COH -+ mov r10, 0 -+ ldm sp++, r5, r6, r7,pc -+ -+__avr32_f64_mul_res_nan: /* Return NaN. */ -+ mov r11, -1 -+ mov r10, -1 -+ ldm sp++, r5, r6, r7,pc -+ -+__avr32_f64_mul_res_inf: /* Return INF. */ -+ mov r11, 0 -+ orh r11, 0x7ff0 -+ bld lr, 31 -+ bst r11, 31 -+ mov r10, 0 -+ ldm sp++, r5, r6, r7,pc -+ -+__avr32_f64_mul_op1_zero: -+ /* Get sign */ -+ eor r11, r11, r9 -+ andh r11, 0x8000, COH -+ /* Check if op2 is Inf or NaN. */ -+ bfextu r12, r9, 20, 11 -+ cp.w r12, 0x7ff -+ retne r12 /* Return 0.0 */ -+ /* Return NaN */ -+ mov r10, -1 -+ mov r11, -1 -+ ret r12 -+ -+ -+ -+#endif -+ -+ -+#if defined(L_avr32_f64_addsub) || defined(L_avr32_f64_addsub_fast) -+ .align 2 -+ -+__avr32_f64_sub_from_add: -+ /* Switch sign on op2 */ -+ eorh r9, 0x8000 -+ -+#if defined(L_avr32_f64_addsub_fast) -+ .global __avr32_f64_sub_fast -+ .type __avr32_f64_sub_fast,@function -+__avr32_f64_sub_fast: -+#else -+ .global __avr32_f64_sub -+ .type __avr32_f64_sub,@function -+__avr32_f64_sub: -+#endif -+ -+ /* op1 in {r11,r10}*/ -+ /* op2 in {r9,r8}*/ -+ -+#if defined(L_avr32_f64_addsub_fast) -+ /* If op2 is zero just return op1 */ -+ or r12, r8, r9 << 1 -+ reteq r12 -+#endif -+ -+ /* Check signs */ -+ eor r12, r11, r9 -+ /* Different signs, use addition. */ -+ brmi __avr32_f64_add_from_sub -+ -+ stm --sp, r5, r6, r7, lr -+ -+ /* Get sign of op1 into r12 */ -+ mov r12, r11 -+ andh r12, 0x8000, COH -+ -+ /* Remove sign from operands */ -+ cbr r11, 31 -+ cbr r9, 31 -+ -+ /* Put the number with the largest exponent in [r11, r10] -+ and the number with the smallest exponent in [r9, r8] */ -+ cp r11, r9 -+ brhs 1f /* Skip swap if operands already correctly ordered*/ -+ /* Operands were not correctly ordered, swap them*/ -+ mov r7, r11 -+ mov r11, r9 -+ mov r9, r7 -+ mov r7, r10 -+ mov r10, r8 -+ mov r8, r7 -+ eorh r12, 0x8000 /* Invert sign in r12*/ -+1: -+ /* Unpack largest operand - opH */ -+ /* exp: r7 */ -+ /* sf: r11, r10 */ -+ lsr r7, r11, 20 /* Extract exponent */ -+ cbr r7, 11 /* Clear sign bit */ -+ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r11, r11, r10>>21 -+ lsl r10, 11 -+ sbr r11, 31 /* Insert implicit bit */ -+ -+ -+ /* Unpack smallest operand - opL */ -+ /* exp: r6 */ -+ /* sf: r9, r8 */ -+ lsr r6, r9, 20 /* Extract exponent */ -+ cbr r6, 11 /* Clear sign bit */ -+ breq __avr32_f64_sub_opL_subnormal /* If either zero or subnormal */ -+ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>21 -+ lsl r8, 11 -+ sbr r9, 31 /* Insert implicit bit */ -+ -+ -+__avr32_f64_sub_opL_subnormal_done: -+ /* opH is NaN or Inf. */ -+ cp.w r7, 0x7ff -+ breq __avr32_f64_sub_opH_nan_or_inf -+ -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r6, r7 -+ breq __avr32_f64_sub_shift_done /* No need to shift, exponents are equal*/ -+ -+ /* Scale mantissa [r9, r8] with amount [r6]. -+ Uses scratch registers [r5] and [lr]. -+ In IEEE mode:Must not forget the sticky bits we intend to shift out. */ -+ -+ rsub r5,r6,32 /* get (32 - shift count) -+ (if shift count > 32 we get a -+ negative value, but that will -+ work as well in the code below.) */ -+ -+ cp.w r6,32 /* handle shifts >= 32 separately */ -+ brhs __avr32_f64_sub_longshift -+ -+ /* small (<32) shift amount, both words are part of the shift -+ first remember whether part that is lost contains any 1 bits ... */ -+ lsl lr,r8,r5 /* shift away bits that are part of -+ final mantissa. only part that goes -+ to lr are bits that will be lost */ -+ -+ /* ... and now to the actual shift */ -+ lsl r5,r9,r5 /* get bits from msw destined for lsw*/ -+ lsr r8,r8,r6 /* shift down lsw of mantissa */ -+ lsr r9,r9,r6 /* shift down msw of mantissa */ -+ or r8,r5 /* combine these bits with prepared lsw*/ -+#if defined(L_avr32_f64_addsub) -+ cp.w lr,0 /* if any '1' bit in part we lost ...*/ -+ srne lr -+ or r8, lr /* ... we need to set sticky bit*/ -+#endif -+ -+__avr32_f64_sub_shift_done: -+ /* Now subtract the mantissas. */ -+ sub r10, r8 -+ sbc r11, r11, r9 -+ -+ /* Normalize the exponent and mantissa pair stored in -+ [r11,r10] and exponent in [r7]. Needs two scratch registers [r6] and [lr]. */ -+ clz r6,r11 /* Check if we have zeros in high bits */ -+ breq __avr32_f64_sub_longnormalize_done /* No need for scaling if no zeros in high bits */ -+ brcs __avr32_f64_sub_longnormalize -+ -+ /* shift amount is smaller than 32, and involves both msw and lsw*/ -+ rsub lr,r6,32 /* shift mantissa */ -+ lsl r11,r11,r6 -+ lsr lr,r10,lr -+ or r11,lr -+ lsl r10,r10,r6 -+ sub r7,r6 /* adjust exponent */ -+ brle __avr32_f64_sub_subnormal_result -+ -+__avr32_f64_sub_longnormalize_done: -+ -+#if defined(L_avr32_f64_addsub) -+ /* Insert the bits we will remove from the mantissa r9[31:21] */ -+ lsl r9, r10, (32 - 11) -+#else -+ /* Keep the last bit shifted out. */ -+ bfextu r9, r10, 10, 1 -+#endif -+ -+ /* Pack final result*/ -+ /* Input: [r7]:exp, [r11, r10]:mant, [r12]:sign in MSB */ -+ /* Result in [r11,r10] */ -+ /* Insert mantissa */ -+ cbr r11, 31 /*Clear implicit bit*/ -+ lsr r10, 11 -+ or r10, r10, r11<<21 -+ lsr r11, 11 -+ /* Insert exponent and sign bit*/ -+ or r11, r11, r7<<20 -+ bld r12, 31 -+ bst r11, 31 -+ -+ /* Round */ -+__avr32_f64_sub_round: -+#if defined(L_avr32_f64_addsub) -+ mov r7, 0 -+ sbr r7, 31 -+ bld r10, 0 -+ subne r7, -1 -+ -+ cp.w r9, r7 -+ srhs r9 -+#endif -+ add r10, r9 -+ acr r11 -+ -+ /* Return result in [r11,r10] */ -+ ldm sp++, r5, r6, r7,pc -+ -+ -+ -+__avr32_f64_sub_opL_subnormal: -+ /* Extract the of mantissa */ -+ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>21 -+ lsl r8, 11 -+ -+ /* Set exponent to 1 if we do not have a zero. */ -+ or lr, r9, r8 -+ movne r6,1 -+ -+ /* Check if opH is also subnormal. If so, clear implicit bit in r11*/ -+ rsub lr, r7, 0 -+ moveq r7,1 -+ bst r11, 31 -+ -+ /* Check if op1 is zero, if so set exponent to 0. */ -+ or lr, r11, r10 -+ moveq r7,0 -+ -+ rjmp __avr32_f64_sub_opL_subnormal_done -+ -+__avr32_f64_sub_opH_nan_or_inf: -+ /* Check if opH is NaN, if so return NaN */ -+ cbr r11, 31 -+ or lr, r11, r10 -+ brne __avr32_f64_sub_return_nan -+ -+ /* opH is Inf. */ -+ /* Check if opL is Inf. or NaN */ -+ cp.w r6, 0x7ff -+ breq __avr32_f64_sub_opL_nan_or_inf -+ ldm sp++, r5, r6, r7, pc/* opL not Inf or NaN, return opH */ -+__avr32_f64_sub_opL_nan_or_inf: -+ cbr r9, 31 -+ or lr, r9, r8 -+ brne __avr32_f64_sub_return_nan -+ mov r10, 0 /* Generate Inf in r11, r10 */ -+ mov r11, 0 -+ orh r11, 0x7ff0 -+ ldm sp++, r5, r6, r7, pc/* opL Inf, return Inf */ -+__avr32_f64_sub_return_nan: -+ mov r10, -1 /* Generate NaN in r11, r10 */ -+ mov r11, -1 -+ ldm sp++, r5, r6, r7, pc/* opL Inf or NaN, return NaN */ -+ -+ -+__avr32_f64_sub_subnormal_result: -+ /* Just flush subnormals to zero. */ -+ mov r10, 0 -+ mov r11, 0 -+ ldm sp++, r5, r6, r7, pc -+ -+ -+__avr32_f64_sub_longshift: -+ /* large (>=32) shift amount, only lsw will have bits left after shift. -+ note that shift operations will use ((shift count=r6) mod 32) so -+ we do not need to subtract 32 from shift count. */ -+ /* Saturate the shift amount to 63. If the amount -+ is any larger op2 is insignificant. */ -+ satu r6 >> 0, 6 -+#if defined(L_avr32_f64_addsub) -+ /* first remember whether part that is lost contains any 1 bits ... */ -+ lsl lr,r9,r5 /* save all lost bits from msw */ -+ or lr,r8 /* also save lost bits (all) from lsw -+ now lr != 0 if we lose any bits */ -+#endif -+ /* ... and now to the actual shift */ -+ mov r8,r9 /* msw -> lsw (i.e. "shift 32 first")*/ -+ mov r9,0 /* clear msw */ -+ lsr r8,r8,r6 /* make rest of shift inside lsw*/ -+#if defined(L_avr32_f64_addsub) -+ cp.w lr,0 /* if any '1' bit in part we lost ...*/ -+ sreq lr -+ or r8, lr /* ... we need to set sticky bit*/ -+#endif -+ rjmp __avr32_f64_sub_shift_done -+ -+__avr32_f64_sub_longnormalize: -+ /* shift amount is greater than 32 */ -+ clz r6,r10 /* shift mantissa */ -+ /* If the resulting mantissa is zero the result is -+ zero so force exponent to zero. */ -+ movcs r7, 0 -+ movcs r6, 0 -+ subcc r6,-32 -+ mov r11,r10 -+ lsl r11,r11,r6 -+ mov r10,0 -+ sub r7,r6 /* adjust exponent */ -+ brle __avr32_f64_sub_subnormal_result -+ rjmp __avr32_f64_sub_longnormalize_done -+ -+ -+ -+ .align 2 -+__avr32_f64_add_from_sub: -+ /* Switch sign on op2 */ -+ eorh r9, 0x8000 -+ -+#if defined(L_avr32_f64_addsub_fast) -+ .global __avr32_f64_add_fast -+ .type __avr32_f64_add_fast,@function -+__avr32_f64_add_fast: -+#else -+ .global __avr32_f64_add -+ .type __avr32_f64_add,@function -+__avr32_f64_add: -+#endif -+ -+ /* op1 in {r11,r10}*/ -+ /* op2 in {r9,r8}*/ -+ -+#if defined(L_avr32_f64_addsub_fast) -+ /* If op2 is zero just return op1 */ -+ or r12, r8, r9 << 1 -+ reteq r12 -+#endif -+ -+ /* Check signs */ -+ eor r12, r11, r9 -+ /* Different signs, use subtraction. */ -+ brmi __avr32_f64_sub_from_add -+ -+ stm --sp, r5, r6, r7, lr -+ -+ /* Get sign of op1 into r12 */ -+ mov r12, r11 -+ andh r12, 0x8000, COH -+ -+ /* Remove sign from operands */ -+ cbr r11, 31 -+ cbr r9, 31 -+ -+ /* Put the number with the largest exponent in [r11, r10] -+ and the number with the smallest exponent in [r9, r8] */ -+ cp r11, r9 -+ brhs 1f /* Skip swap if operands already correctly ordered*/ -+ /* Operands were not correctly ordered, swap them*/ -+ mov r7, r11 -+ mov r11, r9 -+ mov r9, r7 -+ mov r7, r10 -+ mov r10, r8 -+ mov r8, r7 -+1: -+ /* Unpack largest operand - opH */ -+ /* exp: r7 */ -+ /* sf: r11, r10 */ -+ lsr r7, r11, 20 /* Extract exponent */ -+ cbr r7, 11 /* Clear sign bit */ -+ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r11, r11, r10>>21 -+ lsl r10, 11 -+ sbr r11, 31 /* Insert implicit bit */ -+ -+ -+ /* Unpack smallest operand - opL */ -+ /* exp: r6 */ -+ /* sf: r9, r8 */ -+ lsr r6, r9, 20 /* Extract exponent */ -+ cbr r6, 11 /* Clear sign bit */ -+ breq __avr32_f64_add_opL_subnormal /* If either zero or subnormal */ -+ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>21 -+ lsl r8, 11 -+ sbr r9, 31 /* Insert implicit bit */ -+ -+ -+__avr32_f64_add_opL_subnormal_done: -+ /* opH is NaN or Inf. */ -+ cp.w r7, 0x7ff -+ breq __avr32_f64_add_opH_nan_or_inf -+ -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r6, r7 -+ breq __avr32_f64_add_shift_done /* No need to shift, exponents are equal*/ -+ -+ /* Scale mantissa [r9, r8] with amount [r6]. -+ Uses scratch registers [r5] and [lr]. -+ In IEEE mode:Must not forget the sticky bits we intend to shift out. */ -+ rsub r5,r6,32 /* get (32 - shift count) -+ (if shift count > 32 we get a -+ negative value, but that will -+ work as well in the code below.) */ -+ -+ cp.w r6,32 /* handle shifts >= 32 separately */ -+ brhs __avr32_f64_add_longshift -+ -+ /* small (<32) shift amount, both words are part of the shift -+ first remember whether part that is lost contains any 1 bits ... */ -+ lsl lr,r8,r5 /* shift away bits that are part of -+ final mantissa. only part that goes -+ to lr are bits that will be lost */ -+ -+ /* ... and now to the actual shift */ -+ lsl r5,r9,r5 /* get bits from msw destined for lsw*/ -+ lsr r8,r8,r6 /* shift down lsw of mantissa */ -+ lsr r9,r9,r6 /* shift down msw of mantissa */ -+ or r8,r5 /* combine these bits with prepared lsw*/ -+#if defined(L_avr32_f64_addsub) -+ cp.w lr,0 /* if any '1' bit in part we lost ...*/ -+ srne lr -+ or r8, lr /* ... we need to set sticky bit*/ -+#endif -+ -+__avr32_f64_add_shift_done: -+ /* Now add the mantissas. */ -+ add r10, r8 -+ adc r11, r11, r9 -+ -+ /* Check if we overflowed. */ -+ brcs __avr32_f64_add_res_of -+ -+__avr32_f64_add_res_of_done: -+ -+#if defined(L_avr32_f64_addsub) -+ /* Insert the bits we will remove from the mantissa r9[31:21] */ -+ lsl r9, r10, (32 - 11) -+#else -+ /* Keep the last bit shifted out. */ -+ bfextu r9, r10, 10, 1 -+#endif -+ -+ /* Pack final result*/ -+ /* Input: [r7]:exp, [r11, r10]:mant, [r12]:sign in MSB */ -+ /* Result in [r11,r10] */ -+ /* Insert mantissa */ -+ cbr r11, 31 /*Clear implicit bit*/ -+ lsr r10, 11 -+ or r10, r10, r11<<21 -+ lsr r11, 11 -+ /* Insert exponent and sign bit*/ -+ or r11, r11, r7<<20 -+ bld r12, 31 -+ bst r11, 31 -+ -+ /* Round */ -+__avr32_f64_add_round: -+#if defined(L_avr32_f64_addsub) -+ mov r7, 0 -+ sbr r7, 31 -+ bld r10, 0 -+ subne r7, -1 -+ -+ cp.w r9, r7 -+ srhs r9 -+#endif -+ add r10, r9 -+ acr r11 -+ -+ /* Return result in [r11,r10] */ -+ ldm sp++, r5, r6, r7,pc -+ -+ -+ -+__avr32_f64_add_opL_subnormal: -+ /* Extract the of mantissa */ -+ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>21 -+ lsl r8, 11 -+ -+ /* Set exponent to 1 if we do not have a zero. */ -+ or lr, r9, r8 -+ movne r6,1 -+ -+ /* Check if opH is also subnormal. If so, clear implicit bit in r11*/ -+ rsub lr, r7, 0 -+ moveq r7,1 -+ bst r11, 31 -+ -+ /* Check if op1 is zero, if so set exponent to 0. */ -+ or lr, r11, r10 -+ moveq r7,0 -+ -+ rjmp __avr32_f64_add_opL_subnormal_done -+ -+__avr32_f64_add_opH_nan_or_inf: -+ /* Check if opH is NaN, if so return NaN */ -+ cbr r11, 31 -+ or lr, r11, r10 -+ brne __avr32_f64_add_return_nan -+ -+ /* opH is Inf. */ -+ /* Check if opL is Inf. or NaN */ -+ cp.w r6, 0x7ff -+ breq __avr32_f64_add_opL_nan_or_inf -+ ldm sp++, r5, r6, r7, pc/* opL not Inf or NaN, return opH */ -+__avr32_f64_add_opL_nan_or_inf: -+ cbr r9, 31 -+ or lr, r9, r8 -+ brne __avr32_f64_add_return_nan -+ mov r10, 0 /* Generate Inf in r11, r10 */ -+ mov r11, 0 -+ orh r11, 0x7ff0 -+ ldm sp++, r5, r6, r7, pc/* opL Inf, return Inf */ -+__avr32_f64_add_return_nan: -+ mov r10, -1 /* Generate NaN in r11, r10 */ -+ mov r11, -1 -+ ldm sp++, r5, r6, r7, pc/* opL Inf or NaN, return NaN */ -+ -+ -+__avr32_f64_add_longshift: -+ /* large (>=32) shift amount, only lsw will have bits left after shift. -+ note that shift operations will use ((shift count=r6) mod 32) so -+ we do not need to subtract 32 from shift count. */ -+ /* Saturate the shift amount to 63. If the amount -+ is any larger op2 is insignificant. */ -+ satu r6 >> 0, 6 -+#if defined(L_avr32_f64_addsub) -+ /* first remember whether part that is lost contains any 1 bits ... */ -+ lsl lr,r9,r5 /* save all lost bits from msw */ -+ or lr,r8 /* also save lost bits (all) from lsw -+ now lr != 0 if we lose any bits */ -+#endif -+ /* ... and now to the actual shift */ -+ mov r8,r9 /* msw -> lsw (i.e. "shift 32 first")*/ -+ mov r9,0 /* clear msw */ -+ lsr r8,r8,r6 /* make rest of shift inside lsw*/ -+#if defined(L_avr32_f64_addsub) -+ cp.w lr,0 /* if any '1' bit in part we lost ...*/ -+ sreq lr -+ or r8, lr /* ... we need to set sticky bit*/ -+#endif -+ rjmp __avr32_f64_add_shift_done -+ -+__avr32_f64_add_res_of: -+ /* We overflowed. Increase exponent and shift mantissa.*/ -+ /* [r7]:exp, [r11, r10]:mant */ -+ ror r11 -+ lsr r10, 1 -+ sub r7, -1 -+ -+ /* Clear mantissa to set result to Inf if the exponent is 255. */ -+ cp.w r7, 0x7ff -+ moveq r10, 0 -+ moveq r11, 0 -+ rjmp __avr32_f64_add_res_of_done -+ -+ -+#endif -+ -+#ifdef L_avr32_f64_to_u32 -+ /* This goes into L_fixdfsi */ -+#endif -+ -+ -+#ifdef L_avr32_f64_to_s32 -+ .global __avr32_f64_to_u32 -+ .type __avr32_f64_to_u32,@function -+__avr32_f64_to_u32: -+ cp.w r11, 0 -+ retmi 0 /* Negative returns 0 */ -+ -+ /* Fallthrough to df to signed si conversion */ -+ .global __avr32_f64_to_s32 -+ .type __avr32_f64_to_s32,@function -+__avr32_f64_to_s32: -+ lsl r12,r11,1 -+ lsr r12,21 /* extract exponent*/ -+ sub r12,1023 /* convert to unbiased exponent.*/ -+ retlo 0 /* too small exponent implies zero. */ -+ -+1: -+ rsub r12,r12,31 /* shift count = 31 - exponent */ -+ mov r9,r11 /* save sign for later...*/ -+ lsl r11,11 /* remove exponent and sign*/ -+ sbr r11,31 /* add implicit bit*/ -+ or r11,r11,r10>>21 /* get rest of bits from lsw of double */ -+ lsr r11,r11,r12 /* shift down mantissa to final place */ -+ lsl r9,1 /* sign -> carry */ -+ retcc r11 /* if positive, we are done */ -+ neg r11 /* if negative float, negate result */ -+ ret r11 -+ -+#endif /* L_fixdfsi*/ -+ -+#ifdef L_avr32_f64_to_u64 -+ /* Actual function is in L_fixdfdi */ -+#endif -+ -+#ifdef L_avr32_f64_to_s64 -+ .global __avr32_f64_to_u64 -+ .type __avr32_f64_to_u64,@function -+__avr32_f64_to_u64: -+ cp.w r11,0 -+ /* Negative numbers return zero */ -+ movmi r10, 0 -+ movmi r11, 0 -+ retmi r11 -+ -+ -+ -+ /* Fallthrough */ -+ .global __avr32_f64_to_s64 -+ .type __avr32_f64_to_s64,@function -+__avr32_f64_to_s64: -+ lsl r9,r11,1 -+ lsr r9,21 /* get exponent*/ -+ sub r9,1023 /* convert to correct range*/ -+ /* Return zero if exponent to small */ -+ movlo r10, 0 -+ movlo r11, 0 -+ retlo r11 -+ -+ mov r8,r11 /* save sign for later...*/ -+1: -+ lsl r11,11 /* remove exponent */ -+ sbr r11,31 /* add implicit bit*/ -+ or r11,r11,r10>>21 /* get rest of bits from lsw of double*/ -+ lsl r10,11 /* align lsw correctly as well */ -+ rsub r9,r9,63 /* shift count = 63 - exponent */ -+ breq 1f -+ -+ cp.w r9,32 /* is shift count more than one reg? */ -+ brhs 0f -+ -+ mov r12,r11 /* save msw */ -+ lsr r10,r10,r9 /* small shift count, shift down lsw */ -+ lsr r11,r11,r9 /* small shift count, shift down msw */ -+ rsub r9,r9,32 /* get 32-size of shifted out tail */ -+ lsl r12,r12,r9 /* align part to move from msw to lsw */ -+ or r10,r12 /* combine to get new lsw */ -+ rjmp 1f -+ -+0: -+ lsr r10,r11,r9 /* large shift count,only lsw get bits -+ note that shift count is modulo 32*/ -+ mov r11,0 /* msw will be 0 */ -+ -+1: -+ lsl r8,1 /* sign -> carry */ -+ retcc r11 /* if positive, we are done */ -+ -+ neg r11 /* if negative float, negate result */ -+ neg r10 -+ scr r11 -+ ret r11 -+ -+#endif -+ -+#ifdef L_avr32_u32_to_f64 -+ /* Code located in L_floatsidf */ -+#endif -+ -+#ifdef L_avr32_s32_to_f64 -+ .global __avr32_u32_to_f64 -+ .type __avr32_u32_to_f64,@function -+__avr32_u32_to_f64: -+ sub r11, r12, 0 /* Move to r11 and force Z flag to be updated */ -+ mov r12, 0 /* always positive */ -+ rjmp 0f /* Jump to common code for floatsidf */ -+ -+ .global __avr32_s32_to_f64 -+ .type __avr32_s32_to_f64,@function -+__avr32_s32_to_f64: -+ mov r11, r12 /* Keep original value in r12 for sign */ -+ abs r11 /* Absolute value if r12 */ -+0: -+ mov r10,0 /* let remaining bits be zero */ -+ reteq r11 /* zero long will return zero float */ -+ -+ pushm lr -+ mov r9,31+1023 /* set exponent */ -+ -+ normalize_df r9 /*exp*/, r10, r11 /* mantissa */, r8, lr /* scratch */ -+ -+ /* Check if a subnormal result was created */ -+ cp.w r9, 0 -+ brgt 0f -+ -+ adjust_subnormal_df r9 /* exp */, r10, r11 /* Mantissa */, r12 /*sign*/, r8, lr /* scratch */ -+ popm pc -+0: -+ -+ /* Round result */ -+ round_df r9 /*exp*/, r10, r11 /* Mantissa */, r8 /*scratch*/ -+ cp.w r9,0x7ff -+ brlt 0f -+ /*Return infinity */ -+ mov r10, 0 -+ mov r11, 0 -+ orh r11, 0xffe0 -+ rjmp __floatsidf_return_op1 -+ -+0: -+ -+ /* Pack */ -+ pack_df r9 /*exp*/, r10, r11 /* mantissa */, r10, r11 /* Output df number*/ -+__floatsidf_return_op1: -+ lsl r12,1 /* shift in sign bit */ -+ ror r11 -+ -+ popm pc -+#endif -+ -+ -+#ifdef L_avr32_f32_cmp_eq -+ .global __avr32_f32_cmp_eq -+ .type __avr32_f32_cmp_eq,@function -+__avr32_f32_cmp_eq: -+ cp.w r12, r11 -+ breq 0f -+ /* If not equal check for +/-0 */ -+ /* Or together the two values and shift out the sign bit. -+ If the result is zero, then the two values are both zero. */ -+ or r12, r11 -+ lsl r12, 1 -+ sreq r12 -+ ret r12 -+0: -+ /* Numbers were equal. Check for NaN or Inf */ -+ mov r11, 0 -+ orh r11, 0xff00 -+ lsl r12, 1 -+ cp.w r12, r11 -+ srls r12 /* 0 if NaN, 1 otherwise */ -+ ret r12 -+#endif -+ -+#if defined(L_avr32_f32_cmp_ge) || defined(L_avr32_f32_cmp_lt) -+#ifdef L_avr32_f32_cmp_ge -+ .global __avr32_f32_cmp_ge -+ .type __avr32_f32_cmp_ge,@function -+__avr32_f32_cmp_ge: -+#endif -+#ifdef L_avr32_f32_cmp_lt -+ .global __avr32_f32_cmp_lt -+ .type __avr32_f32_cmp_lt,@function -+__avr32_f32_cmp_lt: -+#endif -+ lsl r10, r12, 1 /* Remove sign bits */ -+ lsl r9, r11, 1 -+ mov r8, 0 -+ orh r8, 0xff00 -+ cp.w r10, r8 -+ rethi 0 /* Op0 is NaN */ -+ cp.w r9, r8 -+ rethi 0 /* Op1 is Nan */ -+ -+ eor r8, r11, r12 -+ bld r12, 31 -+#ifdef L_avr32_f32_cmp_ge -+ srcc r8 /* Set result to true if op0 is positive*/ -+#endif -+#ifdef L_avr32_f32_cmp_lt -+ srcs r8 /* Set result to true if op0 is negative*/ -+#endif -+ retmi r8 /* Return if signs are different */ -+ brcs 0f /* Both signs negative? */ -+ -+ /* Both signs positive */ -+ cp.w r12, r11 -+#ifdef L_avr32_f32_cmp_ge -+ srhs r12 -+#endif -+#ifdef L_avr32_f32_cmp_lt -+ srlo r12 -+#endif -+ retal r12 -+0: -+ /* Both signs negative */ -+ cp.w r11, r12 -+#ifdef L_avr32_f32_cmp_ge -+ srhs r12 -+#endif -+#ifdef L_avr32_f32_cmp_lt -+ srlo r12 -+#endif -+ retal r12 -+#endif -+ -+ -+#ifdef L_avr32_f64_cmp_eq -+ .global __avr32_f64_cmp_eq -+ .type __avr32_f64_cmp_eq,@function -+__avr32_f64_cmp_eq: -+ cp.w r10,r8 -+ cpc r11,r9 -+ breq 0f -+ -+ /* Args were not equal*/ -+ /* Both args could be zero with different sign bits */ -+ lsl r11,1 /* get rid of sign bits */ -+ lsl r9,1 -+ or r11,r10 /* Check if all bits are zero */ -+ or r11,r9 -+ or r11,r8 -+ sreq r12 /* If all zeros the arguments are equal -+ so return 1 else return 0 */ -+ ret r12 -+0: -+ /* check for NaN */ -+ lsl r11,1 -+ mov r12, 0 -+ orh r12, 0xffe0 -+ cp.w r10,0 -+ cpc r11,r12 /* check if nan or inf */ -+ srls r12 /* If Arg is NaN return 0 else 1*/ -+ ret r12 /* Return */ -+ -+#endif -+ -+ -+#if defined(L_avr32_f64_cmp_ge) || defined(L_avr32_f64_cmp_lt) -+ -+#ifdef L_avr32_f64_cmp_ge -+ .global __avr32_f64_cmp_ge -+ .type __avr32_f64_cmp_ge,@function -+__avr32_f64_cmp_ge: -+#endif -+#ifdef L_avr32_f64_cmp_lt -+ .global __avr32_f64_cmp_lt -+ .type __avr32_f64_cmp_lt,@function -+__avr32_f64_cmp_lt: -+#endif -+ -+ /* compare magnitude of op1 and op2 */ -+ pushm lr -+ -+ lsl r11,1 /* Remove sign bit of op1 */ -+ srcs lr /* Sign op1 to lsb of lr*/ -+ lsl r9,1 /* Remove sign bit of op2 */ -+ rol lr /* Sign op2 to lsb of lr, sign bit op1 bit 1 of lr*/ -+ -+ /* Check for Nan */ -+ mov r12, 0 -+ orh r12, 0xffe0 -+ cp.w r10,0 -+ cpc r11,r12 -+ movhi r12, 0 /* Return false for NaN */ -+ brhi 0f /* We have NaN */ -+ cp.w r8,0 -+ cpc r9,r12 -+ movhi r12, 0 /* Return false for NaN */ -+ brhi 0f /* We have NaN */ -+ -+ cp.w lr,3 /* both operands negative ?*/ -+ breq 1f -+ -+ cp.w lr,1 /* both operands positive? */ -+ brlo 2f -+ -+ /* Different signs. If sign of op1 is negative the difference -+ between op1 and op2 will always be negative, and if op1 is -+ positive the difference will always be positive */ -+#ifdef L_avr32_f64_cmp_ge -+ sreq r12 -+#endif -+#ifdef L_avr32_f64_cmp_lt -+ srne r12 -+#endif -+ popm pc -+ -+ -+2: -+ /* Both operands positive. Just compute the difference */ -+ cp.w r10,r8 -+ cpc r11,r9 -+#ifdef L_avr32_f64_cmp_ge -+ srhs r12 -+#endif -+#ifdef L_avr32_f64_cmp_lt -+ srlo r12 -+#endif -+ popm pc -+ -+1: -+ /* Both operands negative. Compute the difference with operands switched */ -+ cp r8,r10 -+ cpc r9,r11 -+#ifdef L_avr32_f64_cmp_ge -+ srhs r12 -+#endif -+#ifdef L_avr32_f64_cmp_lt -+ srlo r12 -+#endif -+0: -+ popm pc -+#endif -+ -+ -+ -+#ifdef L_avr32_f64_div -+ .align 2 -+ .global __avr32_f64_div -+ .type __avr32_f64_div,@function -+ -+__avr32_f64_div: -+ stm --sp, r0, r1, r2, r3, r4, r5, r6, r7,lr -+ /* op1 in {r11,r10}*/ -+ /* op2 in {r9,r8}*/ -+ eor lr, r11, r9 /* MSB(lr) = Sign(op1) ^ Sign(op2) */ -+ -+ -+ /* Unpack op1 to 2.62 format*/ -+ /* exp: r7 */ -+ /* sf: r11, r10 */ -+ lsr r7, r11, 20 /* Extract exponent */ -+ -+ lsl r11, 9 /* Extract mantissa, leave room for implicit bit */ -+ or r11, r11, r10>>23 -+ lsl r10, 9 -+ sbr r11, 29 /* Insert implicit bit */ -+ andh r11, 0x3fff /*Mask last part of exponent since we use 2.62 format*/ -+ -+ cbr r7, 11 /* Clear sign bit */ -+ /* Check if normalization is needed */ -+ breq 11f /*If number is subnormal, normalize it */ -+22: -+ cp r7, 0x7ff -+ brhs 2f /* Check op1 for NaN or Inf */ -+ -+ /* Unpack op2 to 2.62 format*/ -+ /* exp: r6 */ -+ /* sf: r9, r8 */ -+ lsr r6, r9, 20 /* Extract exponent */ -+ -+ lsl r9, 9 /* Extract mantissa, leave room for implicit bit */ -+ or r9, r9, r8>>23 -+ lsl r8, 9 -+ sbr r9, 29 /* Insert implicit bit */ -+ andh r9, 0x3fff /*Mask last part of exponent since we use 2.62 format*/ -+ -+ cbr r6, 11 /* Clear sign bit */ -+ /* Check if normalization is needed */ -+ breq 13f /*If number is subnormal, normalize it */ -+23: -+ cp r6, 0x7ff -+ brhs 3f /* Check op2 for NaN or Inf */ -+ -+ /* Calculate new exponent */ -+ sub r7, r6 -+ sub r7,-1023 -+ -+ /* Divide */ -+ /* Approximating 1/d with the following recurrence: */ -+ /* R[j+1] = R[j]*(2-R[j]*d) */ -+ /* Using 2.62 format */ -+ /* TWO: r12 */ -+ /* d = op2 = divisor (2.62 format): r9,r8 */ -+ /* Multiply result : r5, r4 */ -+ /* Initial guess : r3, r2 */ -+ /* New approximations : r3, r2 */ -+ /* op1 = Dividend (2.62 format) : r11, r10 */ -+ -+ mov r12, 1 /* Load TWO */ -+ brev r12 -+ -+ /* Load initial guess, using look-up table */ -+ /* Initial guess is of format 01.XY, where XY is constructed as follows: */ -+ /* Let d be of following format: 00.1xy....., then XY=~xy */ -+ /* For d=00.100 = 0,5 -> initial guess=01.11 = 1,75 */ -+ /* For d=00.101 = 0,625 -> initial guess=01.11 = 1,5 */ -+ /* For d=00.110 = 0,75 -> initial guess=01.11 = 1,25 */ -+ /* For d=00.111 = 0,875 -> initial guess=01.11 = 1,0 */ -+ /* r2 is also part of the reg pair forming initial guess, but it*/ -+ /* is kept uninitialized to save one cycle since it has so low significance*/ -+ -+ lsr r3, r12, 1 -+ bfextu r4, r9, 27, 2 -+ com r4 -+ bfins r3, r4, 28, 2 -+ -+ /* First approximation */ -+ /* Approximating to 32 bits */ -+ /* r5 = R[j]*d */ -+ mulu.d r4, r3, r9 -+ /* r5 = 2-R[j]*d */ -+ sub r5, r12, r5<<2 -+ /* r3 = R[j]*(2-R[j]*d) */ -+ mulu.d r4, r3, r5 -+ lsl r3, r5, 2 -+ -+ /* Second approximation */ -+ /* Approximating to 32 bits */ -+ /* r5 = R[j]*d */ -+ mulu.d r4, r3, r9 -+ /* r5 = 2-R[j]*d */ -+ sub r5, r12, r5<<2 -+ /* r3 = R[j]*(2-R[j]*d) */ -+ mulu.d r4, r3, r5 -+ lsl r3, r5, 2 -+ -+ /* Third approximation */ -+ /* Approximating to 32 bits */ -+ /* r5 = R[j]*d */ -+ mulu.d r4, r3, r9 -+ /* r5 = 2-R[j]*d */ -+ sub r5, r12, r5<<2 -+ /* r3 = R[j]*(2-R[j]*d) */ -+ mulu.d r4, r3, r5 -+ lsl r3, r5, 2 -+ -+ /* Fourth approximation */ -+ /* Approximating to 64 bits */ -+ /* r5,r4 = R[j]*d */ -+ mul_approx_df r3 /*ah*/, r2 /*al*/, r9 /*bh*/, r8 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ -+ lsl r5, 2 -+ or r5, r5, r4>>30 -+ lsl r4, 2 -+ /* r5,r4 = 2-R[j]*d */ -+ neg r4 -+ sbc r5, r12, r5 -+ /* r3,r2 = R[j]*(2-R[j]*d) */ -+ mul_approx_df r3 /*ah*/, r2 /*al*/, r5 /*bh*/, r4 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ -+ lsl r3, r5, 2 -+ or r3, r3, r4>>30 -+ lsl r2, r4, 2 -+ -+ -+ /* Fifth approximation */ -+ /* Approximating to 64 bits */ -+ /* r5,r4 = R[j]*d */ -+ mul_approx_df r3 /*ah*/, r2 /*al*/, r9 /*bh*/, r8 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ -+ lsl r5, 2 -+ or r5, r5, r4>>30 -+ lsl r4, 2 -+ /* r5,r4 = 2-R[j]*d */ -+ neg r4 -+ sbc r5, r12, r5 -+ /* r3,r2 = R[j]*(2-R[j]*d) */ -+ mul_approx_df r3 /*ah*/, r2 /*al*/, r5 /*bh*/, r4 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ -+ lsl r3, r5, 2 -+ or r3, r3, r4>>30 -+ lsl r2, r4, 2 -+ -+ -+ /* Multiply with dividend to get quotient */ -+ mul_approx_df r3 /*ah*/, r2 /*al*/, r11 /*bh*/, r10 /*bl*/, r3 /*rh*/, r2 /*rl*/, r1 /*sh*/, r0 /*sl*/ -+ -+ /* Shift by 3 to get result in 1.63 format, as required by the exponent. */ -+ /* Note that 1.63 format is already used by the exponent in r7, since */ -+ /* a bias of 1023 was added to the result exponent, even though the implicit */ -+ /* bit was inserted. This gives the exponent an additional bias of 1, which */ -+ /* supports 1.63 format. */ -+ lsl r3, r3, 3 -+ or r3, r3, r2>>29 -+ lsl r2, r2, 3 -+ -+ /* To increase speed, this result is not corrected before final rounding.*/ -+ /* This may give a difference to IEEE compliant code of 1 ULP.*/ -+ -+ /* Adjust exponent and mantissa */ -+ /* r7:exp, [r3, r2]:mant, [r12,r11,r10]:scratch*/ -+ /* Mantissa may be of the format 0.xxxx or 1.xxxx. */ -+ /* In the first case, shift one pos to left.*/ -+ sub r10, r7, 1 -+ mov r12, r3 -+ lsl r11, r2, 1 -+ rol r12 -+ bld r3, 31 -+ movne r7, r10 -+ movne r3, r12 -+ movne r2, r11 -+ cp r7, 0 -+ breq 15f /*Result was subnormal. Flush-to-zero and return zero*/ -+ -+ /* Result was not subnormal. Perform rounding. */ -+ /* Note that the tie case (for round-to-even) can not occur in division. */ -+ /* [r7]:exp, [r3, r2]:mant */ -+ /* Mantissa is in 0.64 format. Round by adding 1<<(64-(52+2))=1<<10*/ -+ mov r12, (1<<10) -+ add r2, r12 -+ acr r3 -+ /* Adjust exponent if we overflowed.*/ -+ subcs r7, -1 -+ -+ -+ -+ /* Pack final result*/ -+ /* Input: [r7]:exp, [r3, r2]:mant */ -+ /* Result in [r11,r10] */ -+ /* Insert exponent and sign bit*/ -+ lsl r11, r7, 20 -+ bld lr, 31 -+ bst r11, 31 -+ /* Insert mantissa */ -+ cbr r3, 31 /*Clear implicit bit*/ -+ or r11, r11, r3>>11 -+ lsr r10, r2, 11 -+ or r10, r10, r3<<21 -+ /* Return result in [r11,r10] */ -+ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc -+ -+ -+2: -+ /* Op1 is NaN or inf */ -+ andh r11, 0x000f /* Extract mantissa */ -+ or r11, r10 -+ brne 16f /* Return NaN if op1 is NaN */ -+ /* Op1 is inf check op2 */ -+ lsr r6, r9, 20 /* Extract exponent */ -+ cbr r6, 8 /* Clear sign bit */ -+ cp r6, 0x7ff -+ brne 17f /* Inf/number gives inf, return inf */ -+ rjmp 16f /* The rest gives NaN*/ -+ -+3: -+ /* Op1 is a valid number. Op 2 is NaN or inf */ -+ andh r9, 0x000f /* Extract mantissa */ -+ or r9, r8 -+ brne 16f /* Return NaN if op2 is NaN */ -+ rjmp 15f /* Op2 was inf, return zero*/ -+ -+11: /* Op1 was denormal. Fix it. */ -+ lsl r11, 2 -+ or r11, r11, r10 >> 30 -+ lsl r10, 2 -+ cbr r11, 31 -+ /* Check if op1 is zero. */ -+ or r4, r10, r11 -+ breq __avr32_f64_div_op1_zero -+ normalize_df r7 /*exp*/, r10, r11 /*Mantissa*/, r4, r5 /*scratch*/ -+ lsr r10, 2 -+ or r10, r10, r11 << 30 -+ lsr r11, 2 -+ rjmp 22b -+ -+ -+13: /* Op2 was denormal. Fix it */ -+ lsl r9, 2 -+ or r9, r9, r8 >> 30 -+ lsl r8, 2 -+ cbr r9, 31 -+ /* Check if op2 is zero. */ -+ or r4, r9, r8 -+ breq 17f /* Divisor is zero -> return Inf */ -+ normalize_df r6 /*exp*/, r8, r9 /*Mantissa*/, r4, r5 /*scratch*/ -+ lsr r8, 2 -+ or r8, r8, r9 << 30 -+ lsr r9, 2 -+ rjmp 23b -+ -+ -+15: /* Divide result was subnormal. Return zero. */ -+ mov r11, lr /*Get correct sign*/ -+ andh r11, 0x8000, COH -+ mov r10, 0 -+ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc -+ -+16: /* Return NaN. */ -+ mov r11, -1 -+ mov r10, -1 -+ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc -+ -+17: /* Return INF. */ -+ mov r11, lr /*Get correct sign*/ -+ andh r11, 0x8000, COH -+ orh r11, 0x7ff0 -+ mov r10, 0 -+ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc -+ -+__avr32_f64_div_op1_zero: -+ or r5, r8, r9 << 1 -+ breq 16b /* 0.0/0.0 -> NaN */ -+ bfextu r4, r9, 20, 11 -+ cp r4, 0x7ff -+ brne 15b /* Return zero */ -+ /* Check if divisor is Inf or NaN */ -+ or r5, r8, r9 << 12 -+ breq 15b /* Divisor is inf -> return zero */ -+ rjmp 16b /* Return NaN */ -+ -+ -+ -+ -+#endif -+ -+#if defined(L_avr32_f32_addsub) || defined(L_avr32_f32_addsub_fast) -+ -+ .align 2 -+__avr32_f32_sub_from_add: -+ /* Switch sign on op2 */ -+ eorh r11, 0x8000 -+ -+#if defined(L_avr32_f32_addsub_fast) -+ .global __avr32_f32_sub_fast -+ .type __avr32_f32_sub_fast,@function -+__avr32_f32_sub_fast: -+#else -+ .global __avr32_f32_sub -+ .type __avr32_f32_sub,@function -+__avr32_f32_sub: -+#endif -+ -+ /* Check signs */ -+ eor r8, r11, r12 -+ /* Different signs, use subtraction. */ -+ brmi __avr32_f32_add_from_sub -+ -+ /* Get sign of op1 */ -+ mov r8, r12 -+ andh r12, 0x8000, COH -+ -+ /* Remove sign from operands */ -+ cbr r11, 31 -+#if defined(L_avr32_f32_addsub_fast) -+ reteq r8 /* If op2 is zero return op1 */ -+#endif -+ cbr r8, 31 -+ -+ /* Put the number with the largest exponent in r10 -+ and the number with the smallest exponent in r9 */ -+ max r10, r8, r11 -+ min r9, r8, r11 -+ cp r10, r8 /*If largest operand (in R10) is not equal to op1*/ -+ subne r12, 1 /* Subtract 1 from sign, which will invert MSB of r12*/ -+ andh r12, 0x8000, COH /*Mask all but MSB*/ -+ -+ /* Unpack exponent and mantissa of op1 */ -+ lsl r8, r10, 8 -+ sbr r8, 31 /* Set implicit bit. */ -+ lsr r10, 23 -+ -+ /* op1 is NaN or Inf. */ -+ cp.w r10, 0xff -+ breq __avr32_f32_sub_op1_nan_or_inf -+ -+ /* Unpack exponent and mantissa of op2 */ -+ lsl r11, r9, 8 -+ sbr r11, 31 /* Set implicit bit. */ -+ lsr r9, 23 -+ -+#if defined(L_avr32_f32_addsub) -+ /* Keep sticky bit for correct IEEE rounding */ -+ st.w --sp, r12 -+ -+ /* op2 is either zero or subnormal. */ -+ breq __avr32_f32_sub_op2_subnormal -+0: -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r9, r10 -+ -+ /* Saturate the shift amount to 31. If the amount -+ is any larger op2 is insignificant. */ -+ satu r9 >> 0, 5 -+ -+ /* Shift mantissa of op2 to same decimal point as the mantissa -+ of op1. */ -+ lsr r12, r11, r9 -+ -+ /* Put the remainding bits into r11[23:..].*/ -+ rsub r9, r9, (32-8) -+ lsl r11, r11, r9 -+ -+ /* Now subtract the mantissas. */ -+ sub r8, r12 -+ -+ ld.w r12, sp++ -+ -+ /* Normalize resulting mantissa. */ -+ clz r9, r8 -+ lsl r8, r8, r9 -+ sub r10, r9 -+ brle __avr32_f32_sub_subnormal_result -+ -+ /* Insert the bits we will remove from the mantissa into r11[31:24] */ -+ bfins r11, r8, 24, 8 -+#else -+ /* Ignore sticky bit to simplify and speed up rounding */ -+ /* op2 is either zero or subnormal. */ -+ breq __avr32_f32_sub_op2_subnormal -+0: -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r9, r10 -+ -+ /* Saturate the shift amount to 31. If the amount -+ is any larger op2 is insignificant. */ -+ satu r9 >> 0, 5 -+ -+ /* Shift mantissa of op2 to same decimal point as the mantissa -+ of op1. */ -+ lsr r11, r11, r9 -+ -+ /* Now subtract the mantissas. */ -+ sub r8, r11 -+ -+ /* Normalize resulting mantissa. */ -+ clz r9, r8 -+ lsl r8, r8, r9 -+ sub r10, r9 -+ brle __avr32_f32_sub_subnormal_result -+#endif -+ -+ /* Pack result. */ -+ or r12, r12, r8 >> 8 -+ bfins r12, r10, 23, 8 -+ -+ /* Round */ -+__avr32_f32_sub_round: -+#if defined(L_avr32_f32_addsub) -+ mov r10, 0 -+ sbr r10, 31 -+ bld r12, 0 -+ subne r10, -1 -+ cp.w r11, r10 -+ subhs r12, -1 -+#else -+ bld r8, 7 -+ acr r12 -+#endif -+ -+ ret r12 -+ -+ -+__avr32_f32_sub_op2_subnormal: -+ /* Fix implicit bit and adjust exponent of subnormals. */ -+ cbr r11, 31 -+ /* Set exponent to 1 if we do not have a zero. */ -+ movne r9,1 -+ -+ /* Check if op1 is also subnormal. */ -+ cp.w r10, 0 -+ brne 0b -+ -+ cbr r8, 31 -+ /* If op1 is not zero set exponent to 1. */ -+ movne r10,1 -+ -+ rjmp 0b -+ -+__avr32_f32_sub_op1_nan_or_inf: -+ /* Check if op1 is NaN, if so return NaN */ -+ lsl r11, r8, 1 -+ retne -1 -+ -+ /* op1 is Inf. */ -+ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ -+ -+ /* Check if op2 is Inf. or NaN */ -+ lsr r11, r9, 23 -+ cp.w r11, 0xff -+ retne r12 /* op2 not Inf or NaN, return op1 */ -+ -+ ret -1 /* op2 Inf or NaN, return NaN */ -+ -+__avr32_f32_sub_subnormal_result: -+ /* Check if the number is so small that -+ it will be represented with zero. */ -+ rsub r10, r10, 9 -+ rsub r9, r10, 32 -+ retcs 0 -+ -+ /* Shift the mantissa into the correct position.*/ -+ lsr r10, r8, r10 -+ /* Add sign bit. */ -+ or r10, r12 -+ -+ /* Put the shifted out bits in the most significant part -+ of r8. */ -+ lsl r8, r8, r9 -+ -+#if defined(L_avr32_f32_addsub) -+ /* Add all the remainder bits used for rounding into r11 */ -+ andh r11, 0x00FF -+ or r11, r8 -+#else -+ lsr r8, 24 -+#endif -+ rjmp __avr32_f32_sub_round -+ -+ -+ .align 2 -+ -+__avr32_f32_add_from_sub: -+ /* Switch sign on op2 */ -+ eorh r11, 0x8000 -+ -+#if defined(L_avr32_f32_addsub_fast) -+ .global __avr32_f32_add_fast -+ .type __avr32_f32_add_fast,@function -+__avr32_f32_add_fast: -+#else -+ .global __avr32_f32_add -+ .type __avr32_f32_add,@function -+__avr32_f32_add: -+#endif -+ -+ /* Check signs */ -+ eor r8, r11, r12 -+ /* Different signs, use subtraction. */ -+ brmi __avr32_f32_sub_from_add -+ -+ /* Get sign of op1 */ -+ mov r8, r12 -+ andh r12, 0x8000, COH -+ -+ /* Remove sign from operands */ -+ cbr r11, 31 -+#if defined(L_avr32_f32_addsub_fast) -+ reteq r8 /* If op2 is zero return op1 */ -+#endif -+ cbr r8, 31 -+ -+ /* Put the number with the largest exponent in r10 -+ and the number with the smallest exponent in r9 */ -+ max r10, r8, r11 -+ min r9, r8, r11 -+ -+ /* Unpack exponent and mantissa of op1 */ -+ lsl r8, r10, 8 -+ sbr r8, 31 /* Set implicit bit. */ -+ lsr r10, 23 -+ -+ /* op1 is NaN or Inf. */ -+ cp.w r10, 0xff -+ breq __avr32_f32_add_op1_nan_or_inf -+ -+ /* Unpack exponent and mantissa of op2 */ -+ lsl r11, r9, 8 -+ sbr r11, 31 /* Set implicit bit. */ -+ lsr r9, 23 -+ -+#if defined(L_avr32_f32_addsub) -+ /* Keep sticky bit for correct IEEE rounding */ -+ st.w --sp, r12 -+ -+ /* op2 is either zero or subnormal. */ -+ breq __avr32_f32_add_op2_subnormal -+0: -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r9, r10 -+ -+ /* Saturate the shift amount to 31. If the amount -+ is any larger op2 is insignificant. */ -+ satu r9 >> 0, 5 -+ -+ /* Shift mantissa of op2 to same decimal point as the mantissa -+ of op1. */ -+ lsr r12, r11, r9 -+ -+ /* Put the remainding bits into r11[23:..].*/ -+ rsub r9, r9, (32-8) -+ lsl r11, r11, r9 -+ /* Insert the bits we will remove from the mantissa into r11[31:24] */ -+ bfins r11, r12, 24, 8 -+ -+ /* Now add the mantissas. */ -+ add r8, r12 -+ -+ ld.w r12, sp++ -+#else -+ /* Ignore sticky bit to simplify and speed up rounding */ -+ /* op2 is either zero or subnormal. */ -+ breq __avr32_f32_add_op2_subnormal -+0: -+ /* Get shift amount to scale mantissa of op2. */ -+ rsub r9, r10 -+ -+ /* Saturate the shift amount to 31. If the amount -+ is any larger op2 is insignificant. */ -+ satu r9 >> 0, 5 -+ -+ /* Shift mantissa of op2 to same decimal point as the mantissa -+ of op1. */ -+ lsr r11, r11, r9 -+ -+ /* Now add the mantissas. */ -+ add r8, r11 -+ -+#endif -+ /* Check if we overflowed. */ -+ brcs __avr32_f32_add_res_of -+1: -+ /* Pack result. */ -+ or r12, r12, r8 >> 8 -+ bfins r12, r10, 23, 8 -+ -+ /* Round */ -+#if defined(L_avr32_f32_addsub) -+ mov r10, 0 -+ sbr r10, 31 -+ bld r12, 0 -+ subne r10, -1 -+ cp.w r11, r10 -+ subhs r12, -1 -+#else -+ bld r8, 7 -+ acr r12 -+#endif -+ -+ ret r12 -+ -+__avr32_f32_add_op2_subnormal: -+ /* Fix implicit bit and adjust exponent of subnormals. */ -+ cbr r11, 31 -+ /* Set exponent to 1 if we do not have a zero. */ -+ movne r9,1 -+ -+ /* Check if op1 is also subnormal. */ -+ cp.w r10, 0 -+ brne 0b -+ -+ cbr r8, 31 -+ /* If op1 is not zero set exponent to 1. */ -+ movne r10,1 -+ -+ rjmp 0b -+ -+__avr32_f32_add_op1_nan_or_inf: -+ /* Check if op1 is NaN, if so return NaN */ -+ lsl r11, r8, 1 -+ retne -1 -+ -+ /* op1 is Inf. */ -+ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ -+ -+ /* Check if op2 is Inf. or NaN */ -+ lsr r11, r9, 23 -+ cp.w r11, 0xff -+ retne r12 /* op2 not Inf or NaN, return op1 */ -+ -+ lsl r9, 9 -+ reteq r12 /* op2 Inf return op1 */ -+ ret -1 /* op2 is NaN, return NaN */ -+ -+__avr32_f32_add_res_of: -+ /* We overflowed. Increase exponent and shift mantissa.*/ -+ lsr r8, 1 -+ sub r10, -1 -+ -+ /* Clear mantissa to set result to Inf if the exponent is 255. */ -+ cp.w r10, 255 -+ moveq r8, 0 -+ moveq r11, 0 -+ rjmp 1b -+ -+ -+#endif -+ -+ -+#if defined(L_avr32_f32_div) || defined(L_avr32_f32_div_fast) -+ .align 2 -+ -+#if defined(L_avr32_f32_div_fast) -+ .global __avr32_f32_div_fast -+ .type __avr32_f32_div_fast,@function -+__avr32_f32_div_fast: -+#else -+ .global __avr32_f32_div -+ .type __avr32_f32_div,@function -+__avr32_f32_div: -+#endif -+ -+ eor r8, r11, r12 /* MSB(r8) = Sign(op1) ^ Sign(op2) */ -+ -+ /* Unpack */ -+ lsl r12,1 -+ reteq 0 /* Return zero if op1 is zero */ -+ lsl r11,1 -+ breq 4f /* Check op2 for zero */ -+ -+ /* Unpack op1*/ -+ /* exp: r9 */ -+ /* sf: r12 */ -+ lsr r9, r12, 24 -+ breq 11f /*If number is subnormal*/ -+ cp r9, 0xff -+ brhs 2f /* Check op1 for NaN or Inf */ -+ lsl r12, 7 -+ sbr r12, 31 /*Implicit bit*/ -+12: -+ -+ /* Unpack op2*/ -+ /* exp: r10 */ -+ /* sf: r11 */ -+ lsr r10, r11, 24 -+ breq 13f /*If number is subnormal*/ -+ cp r10, 0xff -+ brhs 3f /* Check op2 for NaN or Inf */ -+ -+ lsl r11,7 -+ sbr r11, 31 /*Implicit bit*/ -+14: -+ -+ /* For UC3, store with predecrement is faster than stm */ -+ st.w --sp, r5 -+ st.d --sp, r6 -+ -+ /* Calculate new exponent */ -+ sub r9, r10 -+ sub r9,-127 -+ -+ /* Divide */ -+ /* Approximating 1/d with the following recurrence: */ -+ /* R[j+1] = R[j]*(2-R[j]*d) */ -+ /* Using 2.30 format */ -+ /* TWO: r10 */ -+ /* d: r5 */ -+ /* Multiply result : r6, r7 */ -+ /* Initial guess : r11 */ -+ /* New approximations : r11 */ -+ /* Dividend : r12 */ -+ -+ mov r10, 1 /* Load TWO */ -+ brev r10 -+ -+ lsr r12, 2 /* Get significand of Op1 in 2.30 format */ -+ lsr r5, r11, 2 /* Get significand of Op2 (=d) in 2.30 format */ -+ -+ /* Load initial guess, using look-up table */ -+ /* Initial guess is of format 01.XY, where XY is constructed as follows: */ -+ /* Let d be of following format: 00.1xy....., then XY=~xy */ -+ /* For d=00.100 = 0,5 -> initial guess=01.11 = 1,75 */ -+ /* For d=00.101 = 0,625 -> initial guess=01.11 = 1,5 */ -+ /* For d=00.110 = 0,75 -> initial guess=01.11 = 1,25 */ -+ /* For d=00.111 = 0,875 -> initial guess=01.11 = 1,0 */ -+ -+ lsr r11, r10, 1 -+ bfextu r6, r5, 27, 2 -+ com r6 -+ bfins r11, r6, 28, 2 -+ -+ /* First approximation */ -+ /* r7 = R[j]*d */ -+ mulu.d r6, r11, r5 -+ /* r7 = 2-R[j]*d */ -+ sub r7, r10, r7<<2 -+ /* r11 = R[j]*(2-R[j]*d) */ -+ mulu.d r6, r11, r7 -+ lsl r11, r7, 2 -+ -+ /* Second approximation */ -+ /* r7 = R[j]*d */ -+ mulu.d r6, r11, r5 -+ /* r7 = 2-R[j]*d */ -+ sub r7, r10, r7<<2 -+ /* r11 = R[j]*(2-R[j]*d) */ -+ mulu.d r6, r11, r7 -+ lsl r11, r7, 2 -+ -+ /* Third approximation */ -+ /* r7 = R[j]*d */ -+ mulu.d r6, r11, r5 -+ /* r7 = 2-R[j]*d */ -+ sub r7, r10, r7<<2 -+ /* r11 = R[j]*(2-R[j]*d) */ -+ mulu.d r6, r11, r7 -+ lsl r11, r7, 2 -+ -+ /* Fourth approximation */ -+ /* r7 = R[j]*d */ -+ mulu.d r6, r11, r5 -+ /* r7 = 2-R[j]*d */ -+ sub r7, r10, r7<<2 -+ /* r11 = R[j]*(2-R[j]*d) */ -+ mulu.d r6, r11, r7 -+ lsl r11, r7, 2 -+ -+ -+ /* Multiply with dividend to get quotient, r7 = sf(op1)/sf(op2) */ -+ mulu.d r6, r11, r12 -+ -+ /* Shift by 3 to get result in 1.31 format, as required by the exponent. */ -+ /* Note that 1.31 format is already used by the exponent in r9, since */ -+ /* a bias of 127 was added to the result exponent, even though the implicit */ -+ /* bit was inserted. This gives the exponent an additional bias of 1, which */ -+ /* supports 1.31 format. */ -+ lsl r10, r7, 3 -+ -+#if defined(L_avr32_f32_div) -+ /* To perform correct rounding, check for nonzero remainder, */ -+ /* and set LSB in quot if remainder != 0 */ -+ /* Remainder = dividend(r12) - divisor(r5)*quotient(r10) */ -+ -+ lsl r5, 1 /* Transform divisor from 2.30 to 1.31 format */ -+ /* Mask all bits lower than guard in quotient. */ -+ /* These bits are inexact due to approximative algorithm */ -+ andl r10, 0xffc0 -+ /* Add 1 in least significant bit pos to make sure approximation is from above */ -+ sub r10, -64 -+ mulu.d r6, r5, r10 -+ /* If remainder < 0, truncated quotient is too large, so the */ -+ /* delta added must be subtracted to get the correct truncated quotient. */ -+ sub r12, r7 /* Calculate remainder and implicitly set flags */ -+ sublt r10, 64 -+#endif -+ -+ /* For UC3, load with postincrement is faster than ldm */ -+ ld.d r6, sp++ -+ ld.w r5, sp++ -+ -+ /* Adjust exponent and mantissa */ -+ /* r9:exp, r10:mant, r11:scratch*/ -+ clz r11, r10 -+ sub r9, r11 -+ breq 16f /*Result was subnormal*/ -+ lsl r10, r10, r11 -+ -+ /* Result was not subnormal. Perform rounding. */ -+ /* Note that the tie case (for round-to-even) can not occur in division. */ -+ /* r9:exp, r10:mant*/ -+ sub r10, -1*(0x80) -+ /* Adjust exponent if we overflowed. Note that we must use {cc} -+ since we perform the add using a sub insn. */ -+ subcc r9, -1 -+ -+ /* Pack final result*/ -+ lsr r12, r10, 7 -+ bfins r12, r9, 24, 8 -+__divsf_return_op1: -+ lsl r8, 1 -+ ror r12 -+ ret r12 -+ -+ -+2: -+ /* Op1 is NaN or inf */ -+ retne -1 /* Return NaN if op1 is NaN */ -+ /* Op1 is inf check op2 */ -+ mov r9, 0xff -+ brev r9 -+ cp r11, r9 -+ brlo __divsf_return_op1 /* inf/number gives inf */ -+ ret -1 /* The rest gives NaN*/ -+3: -+ /* Op1 is NaN or inf */ -+ reteq 0 /* Return zero if number/inf*/ -+ ret -1 /* Return NaN*/ -+4: -+ /* Op2 is zero ? */ -+ tst r12,r12 -+ reteq -1 /* 0.0/0.0 is NaN */ -+ /* Nonzero/0.0 is Inf. Sign bit will be shifted in before returning*/ -+ mov r12, 0x7ff -+ brev r12 -+ rjmp __divsf_return_op1 -+ -+11: /* Op1 was denormal. Fix it. */ -+ lsl r12,7 -+ clz r9,r12 -+ lsl r12,r12,r9 -+ rsub r9,r9,1 -+ rjmp 12b -+ -+13: /* Op2 was denormal. Fix it. */ -+ lsl r11,7 -+ clz r10,r11 -+ lsl r11,r11,r10 -+ rsub r10,r10,1 -+ rjmp 14b -+ -+ -+16: /* Divide result was subnormal. Fix it and return. */ -+#if defined(L_avr32_f32_div) -+ lsl r10, r10, r11 /*Perform shift required by adjustment of exponent and mantissa*/ -+ adjust_subnormal_sf r12 /*sf*/, r9 /*exp*/, r10 /*mant*/, r8 /*sign*/,r11 /*scratch*/ -+#else -+ /*Flush to zero*/ -+ mov r12, 0 -+#endif -+ ret r12 -+ -+#endif -+ -+#ifdef L_avr32_f32_mul -+ .global __avr32_f32_mul -+ .type __avr32_f32_mul,@function -+ -+ -+__avr32_f32_mul: -+ mov r8, r12 -+ eor r12, r11 /* MSB(r8) = Sign(op1) ^ Sign(op2) */ -+ andh r12, 0x8000, COH -+ -+ /* arrange operands so that that op1 >= op2 */ -+ cbr r8, 31 -+ breq __avr32_f32_mul_op1_zero -+ cbr r11, 31 -+ -+ /* Put the number with the largest exponent in r10 -+ and the number with the smallest exponent in r9 */ -+ max r10, r8, r11 -+ min r9, r8, r11 -+ -+ /* Unpack exponent and mantissa of op1 */ -+ lsl r8, r10, 8 -+ sbr r8, 31 /* Set implicit bit. */ -+ lsr r10, 23 -+ -+ /* op1 is NaN or Inf. */ -+ cp.w r10, 0xff -+ breq __avr32_f32_mul_op1_nan_or_inf -+ -+ /* Unpack exponent and mantissa of op2 */ -+ lsl r11, r9, 8 -+ sbr r11, 31 /* Set implicit bit. */ -+ lsr r9, 23 -+ -+ /* op2 is either zero or subnormal. */ -+ breq __avr32_f32_mul_op2_subnormal -+0: -+ /* Calculate new exponent */ -+ add r9,r10 -+ -+ /* Do the multiplication */ -+ mulu.d r10,r8,r11 -+ -+ /* We might need to scale up by two if the MSB of the result is -+ zero. */ -+ lsl r8, r11, 1 -+ movcc r11, r8 -+ subcc r9, 1 -+ -+ /* Put the shifted out bits of the mantissa into r10 */ -+ lsr r10, 8 -+ bfins r10, r11, 24, 8 -+ -+ sub r9,(127-1) /* remove extra exponent bias */ -+ brle __avr32_f32_mul_res_subnormal -+ -+ /* Check for Inf. */ -+ cp.w r9, 0xff -+ brge 1f -+ -+ /* Pack result. */ -+ or r12, r12, r11 >> 8 -+ bfins r12, r9, 23, 8 -+ -+ /* Round */ -+__avr32_f32_mul_round: -+ mov r8, 0 -+ sbr r8, 31 -+ bld r12, 0 -+ subne r8, -1 -+ -+ cp.w r10, r8 -+ subhs r12, -1 -+ -+ ret r12 -+ -+1: -+ /* Return Inf */ -+ orh r12, 0x7f80 -+ ret r12 -+ -+__avr32_f32_mul_op2_subnormal: -+ cbr r11, 31 -+ clz r9, r11 -+ retcs 0 /* op2 is zero. Return 0 */ -+ lsl r11, r11, r9 -+ rsub r9, r9, 1 -+ -+ /* Check if op2 is subnormal. */ -+ tst r10, r10 -+ brne 0b -+ -+ /* op2 is subnormal */ -+ cbr r8, 31 -+ clz r10, r11 -+ retcs 0 /* op1 is zero. Return 0 */ -+ lsl r8, r8, r10 -+ rsub r10, r10, 1 -+ -+ rjmp 0b -+ -+ -+__avr32_f32_mul_op1_nan_or_inf: -+ /* Check if op1 is NaN, if so return NaN */ -+ lsl r11, r8, 1 -+ retne -1 -+ -+ /* op1 is Inf. */ -+ tst r9, r9 -+ reteq -1 /* Inf * 0 -> NaN */ -+ -+ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ -+ -+ /* Check if op2 is Inf. or NaN */ -+ lsr r11, r9, 23 -+ cp.w r11, 0xff -+ retne r12 /* op2 not Inf or NaN, return Info */ -+ -+ lsl r9, 9 -+ reteq r12 /* op2 Inf return Inf */ -+ ret -1 /* op2 is NaN, return NaN */ -+ -+__avr32_f32_mul_res_subnormal: -+ /* Check if the number is so small that -+ it will be represented with zero. */ -+ rsub r9, r9, 9 -+ rsub r8, r9, 32 -+ retcs 0 -+ -+ /* Shift the mantissa into the correct position.*/ -+ lsr r9, r11, r9 -+ /* Add sign bit. */ -+ or r12, r9 -+ /* Put the shifted out bits in the most significant part -+ of r8. */ -+ lsl r11, r11, r8 -+ -+ /* Add all the remainder bits used for rounding into r11 */ -+ andh r10, 0x00FF -+ or r10, r11 -+ rjmp __avr32_f32_mul_round -+ -+__avr32_f32_mul_op1_zero: -+ bfextu r10, r11, 23, 8 -+ cp.w r10, 0xff -+ retne r12 -+ reteq -1 -+ -+#endif -+ -+ -+#ifdef L_avr32_s32_to_f32 -+ .global __avr32_s32_to_f32 -+ .type __avr32_s32_to_f32,@function -+__avr32_s32_to_f32: -+ cp r12, 0 -+ reteq r12 /* If zero then return zero float */ -+ mov r11, r12 /* Keep the sign */ -+ abs r12 /* Compute the absolute value */ -+ mov r10, 31 + 127 /* Set the correct exponent */ -+ -+ /* Normalize */ -+ normalize_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ -+ -+ /* Check for subnormal result */ -+ cp.w r10, 0 -+ brle __avr32_s32_to_f32_subnormal -+ -+ round_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ -+ pack_sf r12 /*sf*/, r10 /*exp*/, r12 /*mant*/ -+ lsl r11, 1 -+ ror r12 -+ ret r12 -+ -+__avr32_s32_to_f32_subnormal: -+ /* Adjust a subnormal result */ -+ adjust_subnormal_sf r12/*sf*/, r10 /*exp*/, r12 /*mant*/, r11/*sign*/, r9 /*scratch*/ -+ ret r12 -+ -+#endif -+ -+#ifdef L_avr32_u32_to_f32 -+ .global __avr32_u32_to_f32 -+ .type __avr32_u32_to_f32,@function -+__avr32_u32_to_f32: -+ cp r12, 0 -+ reteq r12 /* If zero then return zero float */ -+ mov r10, 31 + 127 /* Set the correct exponent */ -+ -+ /* Normalize */ -+ normalize_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ -+ -+ /* Check for subnormal result */ -+ cp.w r10, 0 -+ brle __avr32_u32_to_f32_subnormal -+ -+ round_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ -+ pack_sf r12 /*sf*/, r10 /*exp*/, r12 /*mant*/ -+ lsr r12,1 /* Sign bit is 0 for unsigned int */ -+ ret r12 -+ -+__avr32_u32_to_f32_subnormal: -+ /* Adjust a subnormal result */ -+ mov r8, 0 -+ adjust_subnormal_sf r12/*sf*/,r10 /*exp*/, r12 /*mant*/,r8/*sign*/, r9 /*scratch*/ -+ ret r12 -+ -+ -+#endif -+ -+ -+#ifdef L_avr32_f32_to_s32 -+ .global __avr32_f32_to_s32 -+ .type __avr32_f32_to_s32,@function -+__avr32_f32_to_s32: -+ bfextu r11, r12, 23, 8 -+ sub r11,127 /* Fix bias */ -+ retlo 0 /* Negative exponent yields zero integer */ -+ -+ /* Shift mantissa into correct position */ -+ rsub r11,r11,31 /* Shift amount */ -+ lsl r10,r12,8 /* Get mantissa */ -+ sbr r10,31 /* Add implicit bit */ -+ lsr r10,r10,r11 /* Perform shift */ -+ lsl r12,1 /* Check sign */ -+ retcc r10 /* if positive, we are done */ -+ neg r10 /* if negative float, negate result */ -+ ret r10 -+ -+#endif -+ -+#ifdef L_avr32_f32_to_u32 -+ .global __avr32_f32_to_u32 -+ .type __avr32_f32_to_u32,@function -+__avr32_f32_to_u32: -+ cp r12,0 -+ retmi 0 /* Negative numbers gives 0 */ -+ bfextu r11, r12, 23, 8 /* Extract exponent */ -+ sub r11,127 /* Fix bias */ -+ retlo 0 /* Negative exponent yields zero integer */ -+ -+ /* Shift mantissa into correct position */ -+ rsub r11,r11,31 /* Shift amount */ -+ lsl r12,8 /* Get mantissa */ -+ sbr r12,31 /* Add implicit bit */ -+ lsr r12,r12,r11 /* Perform shift */ -+ ret r12 -+ -+#endif -+ -+#ifdef L_avr32_f32_to_f64 -+ .global __avr32_f32_to_f64 -+ .type __avr32_f32_to_f64,@function -+ -+__avr32_f32_to_f64: -+ lsl r11,r12,1 /* Remove sign bit, keep original value in r12*/ -+ moveq r10, 0 -+ reteq r11 /* Return zero if input is zero */ -+ -+ bfextu r9,r11,24,8 /* Get exponent */ -+ cp.w r9,0xff /* check for NaN or inf */ -+ breq 0f -+ -+ lsl r11,7 /* Convert sf mantissa to df format */ -+ mov r10,0 -+ -+ /* Check if implicit bit should be set */ -+ cp.w r9, 0 -+ subeq r9,-1 /* Adjust exponent if it was 0 */ -+ srne r8 -+ or r11, r11, r8 << 31 /* Set implicit bit if needed */ -+ sub r9,(127-0x3ff) /* Convert exponent to df format exponent */ -+ -+ /*We know that low register of mantissa is 0, and will be unaffected by normalization.*/ -+ /*We can therefore use the faster normalize_sf function instead of normalize_df.*/ -+ normalize_sf r9 /*exp*/, r11 /*mantissa*/, r8 /*scratch*/ -+ pack_df r9 /*exp*/, r10, r11 /*mantissa*/, r10, r11 /*df*/ -+ -+__extendsfdf_return_op1: -+ /* Rotate in sign bit */ -+ lsl r12, 1 -+ ror r11 -+ ret r11 -+ -+0: -+ /* Inf or NaN*/ -+ mov r10, 0 -+ orh r10, 0xffe0 -+ lsl r11,8 /* check mantissa */ -+ movne r11, -1 /* Return NaN */ -+ moveq r11, r10 /* Return inf */ -+ rjmp __extendsfdf_return_op1 -+#endif -+ -+ -+#ifdef L_avr32_f64_to_f32 -+ .global __avr32_f64_to_f32 -+ .type __avr32_f64_to_f32,@function -+ -+__avr32_f64_to_f32: -+ /* Unpack */ -+ lsl r9,r11,1 /* Unpack exponent */ -+ lsr r9,21 -+ -+ reteq 0 /* If exponent is 0 the number is so small -+ that the conversion to single float gives -+ zero */ -+ -+ lsl r8,r11,10 /* Adjust mantissa */ -+ or r12,r8,r10>>22 -+ -+ lsl r10,10 /* Check if there are any remaining bits -+ in the low part of the mantissa.*/ -+ neg r10 -+ rol r12 /* If there were remaining bits then set lsb -+ of mantissa to 1 */ -+ -+ cp r9,0x7ff -+ breq 2f /* Check for NaN or inf */ -+ -+ sub r9,(0x3ff-127) /* Adjust bias of exponent */ -+ sbr r12,31 /* set the implicit bit.*/ -+ -+ cp.w r9, 0 /* Check for subnormal number */ -+ brle 3f -+ -+ round_sf r9 /*exp*/, r12 /*mant*/, r10 /*scratch*/ -+ pack_sf r12 /*sf*/, r9 /*exp*/, r12 /*mant*/ -+__truncdfsf_return_op1: -+ /* Rotate in sign bit */ -+ lsl r11, 1 -+ ror r12 -+ ret r12 -+ -+2: -+ /* NaN or inf */ -+ cbr r12,31 /* clear implicit bit */ -+ retne -1 /* Return NaN if mantissa not zero */ -+ mov r12, 0 -+ orh r12, 0xff00 -+ ret r12 /* Return inf */ -+ -+3: /* Result is subnormal. Adjust it.*/ -+ adjust_subnormal_sf r12/*sf*/,r9 /*exp*/, r12 /*mant*/, r11/*sign*/, r10 /*scratch*/ -+ ret r12 -+ -+ -+#endif -+ -\ No newline at end of file -diff -Nrup gcc-4.2.1/gcc/config/avr32/lib2funcs.S gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib2funcs.S ---- gcc-4.2.1/gcc/config/avr32/lib2funcs.S 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib2funcs.S 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,21 @@ -+ .align 4 -+ .global __nonlocal_goto -+ .type __nonlocal_goto,@function -+ -+/* __nonlocal_goto: This function handles nonlocal_goto's in gcc. -+ -+ parameter 0 (r12) = New Frame Pointer -+ parameter 1 (r11) = Address to goto -+ parameter 2 (r10) = New Stack Pointer -+ -+ This function invalidates the return stack, since it returns from a -+ function without using a return instruction. -+*/ -+__nonlocal_goto: -+ mov r7, r12 -+ mov sp, r10 -+ frs # Flush return stack -+ mov pc, r11 -+ -+ -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/linux-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/linux-elf.h ---- gcc-4.2.1/gcc/config/avr32/linux-elf.h 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/linux-elf.h 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,156 @@ -+/* -+ Linux/Elf specific definitions. -+ Copyright 2003-2006 Atmel Corporation. -+ -+ Written by Ronny Pedersen, Atmel Norway, -+ and H�vard Skinnemoen, Atmel Norway, -+ -+ This file is part of GCC. -+ -+ This program is free software; you can redistribute it and/or modify -+ it under the terms of the GNU General Public License as published by -+ the Free Software Foundation; either version 2 of the License, or -+ (at your option) any later version. -+ -+ This program is distributed in the hope that it will be useful, -+ but WITHOUT ANY WARRANTY; without even the implied warranty of -+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ GNU General Public License for more details. -+ -+ You should have received a copy of the GNU General Public License -+ along with this program; if not, write to the Free Software -+ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ -+ -+ -+ -+/* elfos.h should have already been included. Now just override -+ any conflicting definitions and add any extras. */ -+ -+/* Run-time Target Specification. */ -+#undef TARGET_VERSION -+#define TARGET_VERSION fputs (" (AVR32 GNU/Linux with ELF)", stderr); -+ -+/* Do not assume anything about header files. */ -+#define NO_IMPLICIT_EXTERN_C -+ -+/* The GNU C++ standard library requires that these macros be defined. */ -+#undef CPLUSPLUS_CPP_SPEC -+#define CPLUSPLUS_CPP_SPEC "-D_GNU_SOURCE %(cpp)" -+ -+/* Now we define the strings used to build the spec file. */ -+#undef LIB_SPEC -+#define LIB_SPEC \ -+ "%{pthread:-lpthread} \ -+ %{shared:-lc} \ -+ %{!shared:%{profile:-lc_p}%{!profile:-lc}}" -+ -+/* Provide a STARTFILE_SPEC appropriate for GNU/Linux. Here we add -+ the GNU/Linux magical crtbegin.o file (see crtstuff.c) which -+ provides part of the support for getting C++ file-scope static -+ object constructed before entering `main'. */ -+ -+#undef STARTFILE_SPEC -+#define STARTFILE_SPEC \ -+ "%{!shared: \ -+ %{pg:gcrt1.o%s} %{!pg:%{p:gcrt1.o%s} \ -+ %{!p:%{profile:gcrt1.o%s} \ -+ %{!profile:crt1.o%s}}}} \ -+ crti.o%s %{!shared:crtbegin.o%s} %{shared:crtbeginS.o%s}" -+ -+/* Provide a ENDFILE_SPEC appropriate for GNU/Linux. Here we tack on -+ the GNU/Linux magical crtend.o file (see crtstuff.c) which -+ provides part of the support for getting C++ file-scope static -+ object constructed before entering `main', followed by a normal -+ GNU/Linux "finalizer" file, `crtn.o'. */ -+ -+#undef ENDFILE_SPEC -+#define ENDFILE_SPEC \ -+ "%{!shared:crtend.o%s} %{shared:crtendS.o%s} crtn.o%s" -+ -+#undef ASM_SPEC -+#define ASM_SPEC "%{!mno-pic:--pic} %{mrelax|O*:%{mno-relax|O0|O1: ;:--linkrelax}} %{mcpu=*:-mcpu=%*}" -+ -+#undef LINK_SPEC -+#define LINK_SPEC "%{version:-v} \ -+ %{static:-Bstatic} \ -+ %{shared:-shared} \ -+ %{symbolic:-Bsymbolic} \ -+ %{rdynamic:-export-dynamic} \ -+ %{!dynamic-linker:-dynamic-linker /lib/ld-uClibc.so.0} \ -+ %{mrelax|O*:%{mno-relax|O0|O1: ;:--relax}}" -+ -+#define TARGET_OS_CPP_BUILTINS() LINUX_TARGET_OS_CPP_BUILTINS() -+ -+/* This is how we tell the assembler that two symbols have the same value. */ -+#define ASM_OUTPUT_DEF(FILE, NAME1, NAME2) \ -+ do \ -+ { \ -+ assemble_name (FILE, NAME1); \ -+ fputs (" = ", FILE); \ -+ assemble_name (FILE, NAME2); \ -+ fputc ('\n', FILE); \ -+ } \ -+ while (0) -+ -+ -+ -+#undef CC1_SPEC -+#define CC1_SPEC "%{profile:-p}" -+ -+/* Target CPU builtins. */ -+#define TARGET_CPU_CPP_BUILTINS() \ -+ do \ -+ { \ -+ builtin_define ("__avr32__"); \ -+ builtin_define ("__AVR32__"); \ -+ builtin_define ("__AVR32_LINUX__"); \ -+ builtin_define (avr32_part->macro); \ -+ builtin_define (avr32_arch->macro); \ -+ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) \ -+ builtin_define ("__AVR32_AVR32A__"); \ -+ else \ -+ builtin_define ("__AVR32_AVR32B__"); \ -+ if (TARGET_UNALIGNED_WORD) \ -+ builtin_define ("__AVR32_HAS_UNALIGNED_WORD__"); \ -+ if (TARGET_SIMD) \ -+ builtin_define ("__AVR32_HAS_SIMD__"); \ -+ if (TARGET_DSP) \ -+ builtin_define ("__AVR32_HAS_DSP__"); \ -+ if (TARGET_RMW) \ -+ builtin_define ("__AVR32_HAS_RMW__"); \ -+ if (TARGET_BRANCH_PRED) \ -+ builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ -+ if (TARGET_FAST_FLOAT) \ -+ builtin_define ("__AVR32_FAST_FLOAT__"); \ -+ if (flag_pic) \ -+ { \ -+ builtin_define ("__PIC__"); \ -+ builtin_define ("__pic__"); \ -+ } \ -+ } \ -+ while (0) -+ -+ -+ -+/* Call the function profiler with a given profile label. */ -+#undef FUNCTION_PROFILER -+#define FUNCTION_PROFILER(STREAM, LABELNO) \ -+ do \ -+ { \ -+ fprintf (STREAM, "\tmov\tlr, lo(mcount)\n\torh\tlr, hi(mcount)\n"); \ -+ fprintf (STREAM, "\ticall lr\n"); \ -+ } \ -+ while (0) -+ -+#define NO_PROFILE_COUNTERS 1 -+ -+/* For dynamic libraries to work */ -+/* #define PLT_REG_CALL_CLOBBERED 1 */ -+#define AVR32_ALWAYS_PIC 1 -+ -+/* uclibc does not implement sinf, cosf etc. */ -+#undef TARGET_C99_FUNCTIONS -+#define TARGET_C99_FUNCTIONS 0 -+ -+#define LINK_GCC_C_SEQUENCE_SPEC \ -+ "%{static:--start-group} %G %L %{static:--end-group}%{!static:%G}" -diff -Nrup gcc-4.2.1/gcc/config/avr32/predicates.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/predicates.md ---- gcc-4.2.1/gcc/config/avr32/predicates.md 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/predicates.md 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,331 @@ -+;; AVR32 predicates file. -+;; Copyright 2003-2006 Atmel Corporation. -+;; -+;; Written by Ronny Pedersen, Atmel Norway, -+;; -+;; This file is part of GCC. -+;; -+;; This program is free software; you can redistribute it and/or modify -+;; it under the terms of the GNU General Public License as published by -+;; the Free Software Foundation; either version 2 of the License, or -+;; (at your option) any later version. -+;; -+;; This program is distributed in the hope that it will be useful, -+;; but WITHOUT ANY WARRANTY; without even the implied warranty of -+;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+;; GNU General Public License for more details. -+;; -+;; You should have received a copy of the GNU General Public License -+;; along with this program; if not, write to the Free Software -+;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. -+ -+ -+;; True if the operand is a memory reference which contains an -+;; Address consisting of a single pointer register -+(define_predicate "avr32_indirect_register_operand" -+ (and (match_code "mem") -+ (match_test "register_operand(XEXP(op, 0), SImode)"))) -+ -+ -+ -+;; Address expression with a base pointer offset with -+;; a register displacement -+(define_predicate "avr32_indexed_memory_operand" -+ (and (match_code "mem") -+ (match_test "GET_CODE(XEXP(op, 0)) == PLUS")) -+ { -+ -+ rtx op0 = XEXP(XEXP(op, 0), 0); -+ rtx op1 = XEXP(XEXP(op, 0), 1); -+ -+ return ((avr32_address_register_rtx_p (op0, 0) -+ && avr32_legitimate_index_p (GET_MODE(op), op1, 0)) -+ || (avr32_address_register_rtx_p (op1, 0) -+ && avr32_legitimate_index_p (GET_MODE(op), op0, 0))); -+ -+ }) -+ -+;; Operand suitable for the ld.sb instruction -+(define_predicate "load_sb_memory_operand" -+ (ior (match_operand 0 "avr32_indirect_register_operand") -+ (match_operand 0 "avr32_indexed_memory_operand"))) -+ -+ -+;; Operand suitable as operand to insns sign extending QI values -+(define_predicate "extendqi_operand" -+ (ior (match_operand 0 "load_sb_memory_operand") -+ (match_operand 0 "register_operand"))) -+ -+(define_predicate "post_inc_memory_operand" -+ (and (match_code "mem") -+ (match_test "(GET_CODE(XEXP(op, 0)) == POST_INC) -+ && REG_P(XEXP(XEXP(op, 0), 0))"))) -+ -+(define_predicate "pre_dec_memory_operand" -+ (and (match_code "mem") -+ (match_test "(GET_CODE(XEXP(op, 0)) == PRE_DEC) -+ && REG_P(XEXP(XEXP(op, 0), 0))"))) -+ -+;; Operand suitable for add instructions -+(define_predicate "avr32_add_operand" -+ (ior (match_operand 0 "register_operand") -+ (and (match_operand 0 "immediate_operand") -+ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'I', \"Is21\")")))) -+ -+;; Operand is a power of two immediate -+(define_predicate "power_of_two_operand" -+ (match_code "const_int") -+{ -+ HOST_WIDE_INT value = INTVAL (op); -+ -+ return value != 0 && (value & (value - 1)) == 0; -+}) -+ -+;; Operand is a multiple of 8 immediate -+(define_predicate "multiple_of_8_operand" -+ (match_code "const_int") -+{ -+ HOST_WIDE_INT value = INTVAL (op); -+ -+ return (value & 0x7) == 0 ; -+}) -+ -+;; Operand is a multiple of 16 immediate -+(define_predicate "multiple_of_16_operand" -+ (match_code "const_int") -+{ -+ HOST_WIDE_INT value = INTVAL (op); -+ -+ return (value & 0xf) == 0 ; -+}) -+ -+;; Operand is a mask used for masking away upper bits of a reg -+(define_predicate "avr32_mask_upper_bits_operand" -+ (match_code "const_int") -+{ -+ HOST_WIDE_INT value = INTVAL (op) + 1; -+ -+ return value != 1 && value != 0 && (value & (value - 1)) == 0; -+}) -+ -+ -+;; Operand suitable for mul instructions -+(define_predicate "avr32_mul_operand" -+ (ior (match_operand 0 "register_operand") -+ (and (match_operand 0 "immediate_operand") -+ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ks08\")")))) -+ -+;; True for logical binary operators. -+(define_predicate "logical_binary_operator" -+ (match_code "ior,xor,and")) -+ -+;; True for logical shift operators -+(define_predicate "logical_shift_operator" -+ (match_code "ashift,lshiftrt")) -+ -+;; True for shift operand for logical and, or and eor insns -+(define_predicate "avr32_logical_shift_operand" -+ (and (match_code "ashift,lshiftrt") -+ (ior (and (match_test "GET_CODE(XEXP(op, 1)) == CONST_INT") -+ (match_test "register_operand(XEXP(op, 0), GET_MODE(XEXP(op, 0)))")) -+ (and (match_test "GET_CODE(XEXP(op, 0)) == CONST_INT") -+ (match_test "register_operand(XEXP(op, 1), GET_MODE(XEXP(op, 1)))")))) -+ { -+ return 1; -+ } -+ ) -+ -+ -+;; Predicate for second operand to and, ior and xor insn patterns -+(define_predicate "avr32_logical_insn_operand" -+ (ior (match_operand 0 "register_operand") -+ (match_operand 0 "avr32_logical_shift_operand")) -+ { -+ return 1; -+ } -+) -+ -+ -+;; True for avr32 comparison operators -+(define_predicate "avr32_comparison_operator" -+ (ior (match_code "eq, ne, gt, ge, lt, le, gtu, geu, ltu, leu") -+ (and (match_code "unspec") -+ (match_test "(XINT(op, 1) == UNSPEC_COND_MI) -+ || (XINT(op, 1) == UNSPEC_COND_PL)")))) -+ -+;; True for avr32 comparison operand -+(define_predicate "avr32_comparison_operand" -+ (ior (and (match_code "eq, ne, gt, ge, lt, le, gtu, geu, ltu, leu") -+ (match_test "(rtx_equal_p (XEXP(op,0), cc0_rtx) && rtx_equal_p (XEXP(op,1), const0_rtx))")) -+ (and (match_code "unspec") -+ (match_test "(XINT(op, 1) == UNSPEC_COND_MI) -+ || (XINT(op, 1) == UNSPEC_COND_PL)")))) -+ -+;; True if this is a const_int with one bit set -+(define_predicate "one_bit_set_operand" -+ (match_code "const_int") -+ { -+ int i; -+ int value; -+ int ones = 0; -+ -+ value = INTVAL(op); -+ for ( i = 0 ; i < 32; i++ ){ -+ if ( value & ( 1 << i ) ){ -+ ones++; -+ } -+ } -+ -+ return ( ones == 1 ); -+ }) -+ -+ -+;; True if this is a const_int with one bit cleared -+(define_predicate "one_bit_cleared_operand" -+ (match_code "const_int") -+ { -+ int i; -+ int value; -+ int zeroes = 0; -+ -+ value = INTVAL(op); -+ for ( i = 0 ; i < 32; i++ ){ -+ if ( !(value & ( 1 << i )) ){ -+ zeroes++; -+ } -+ } -+ -+ return ( zeroes == 1 ); -+ }) -+ -+ -+;; True if this is a register or immediate operand -+(define_predicate "register_immediate_operand" -+ (ior (match_operand 0 "register_operand") -+ (match_operand 0 "immediate_operand"))) -+ -+ -+;; True is this is an operand containing a label_ref -+(define_predicate "avr32_label_ref_operand" -+ (and (match_code "mem") -+ (match_test "avr32_find_symbol(op) -+ && (GET_CODE(avr32_find_symbol(op)) == LABEL_REF)"))) -+ -+;; True is this is a valid symbol pointing to the constant pool -+(define_predicate "avr32_const_pool_operand" -+ (and (match_code "symbol_ref") -+ (match_test "CONSTANT_POOL_ADDRESS_P(op)")) -+ { -+ return (flag_pic ? (!(symbol_mentioned_p (get_pool_constant (op)) -+ || label_mentioned_p (get_pool_constant (op))) -+ || avr32_got_mentioned_p(get_pool_constant (op))) -+ : true); -+ } -+) -+ -+;; True is this is a memory reference to the constant or mini pool -+(define_predicate "avr32_const_pool_ref_operand" -+ (ior (match_operand 0 "avr32_label_ref_operand") -+ (and (match_code "mem") -+ (match_test "avr32_const_pool_operand(XEXP(op,0), GET_MODE(XEXP(op,0)))")))) -+ -+ -+;; Legal source operand for movti insns -+(define_predicate "avr32_movti_src_operand" -+ (ior (match_operand 0 "avr32_const_pool_ref_operand") -+ (ior (ior (match_operand 0 "register_immediate_operand") -+ (match_operand 0 "avr32_indirect_register_operand")) -+ (match_operand 0 "post_inc_memory_operand")))) -+ -+;; Legal destination operand for movti insns -+(define_predicate "avr32_movti_dst_operand" -+ (ior (ior (match_operand 0 "register_operand") -+ (match_operand 0 "avr32_indirect_register_operand")) -+ (match_operand 0 "pre_dec_memory_operand"))) -+ -+ -+;; True is this is a k12 offseted memory operand -+(define_predicate "avr32_k12_memory_operand" -+ (and (match_code "mem") -+ (ior (match_test "REG_P(XEXP(op, 0))") -+ (match_test "GET_CODE(XEXP(op, 0)) == PLUS -+ && REG_P(XEXP(XEXP(op, 0), 0)) -+ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT) -+ && (CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(XEXP(op, 0), 0)), -+ 'K', (mode == SImode) ? \"Ks14\" : ((mode == HImode) ? \"Ks13\" : \"Ks12\")))")))) -+ -+;; True is this is a memory operand with an immediate displacement -+(define_predicate "avr32_imm_disp_memory_operand" -+ (and (match_code "mem") -+ (match_test "GET_CODE(XEXP(op, 0)) == PLUS -+ && REG_P(XEXP(XEXP(op, 0), 0)) -+ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT)"))) -+ -+;; True is this is a bswap operand -+(define_predicate "avr32_bswap_operand" -+ (ior (match_operand 0 "avr32_k12_memory_operand") -+ (match_operand 0 "register_operand"))) -+ -+;; True is this is a valid coprocessor insn memory operand -+(define_predicate "avr32_cop_memory_operand" -+ (and (match_operand 0 "memory_operand") -+ (not (match_test "GET_CODE(XEXP(op, 0)) == PLUS -+ && REG_P(XEXP(XEXP(op, 0), 0)) -+ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT) -+ && !(CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(XEXP(op, 0), 0)), 'K', \"Ku10\"))")))) -+ -+;; True is this is a valid source/destination operand -+;; for moving values to/from a coprocessor -+(define_predicate "avr32_cop_move_operand" -+ (ior (match_operand 0 "register_operand") -+ (match_operand 0 "avr32_cop_memory_operand"))) -+ -+ -+;; True is this is a valid extract byte offset for use in -+;; load extracted index insns -+(define_predicate "avr32_extract_shift_operand" -+ (and (match_operand 0 "const_int_operand") -+ (match_test "(INTVAL(op) == 0) || (INTVAL(op) == 8) -+ || (INTVAL(op) == 16) || (INTVAL(op) == 24)"))) -+ -+;; True is this is a floating-point register -+(define_predicate "avr32_fp_register_operand" -+ (and (match_operand 0 "register_operand") -+ (match_test "REGNO_REG_CLASS(REGNO(op)) == FP_REGS"))) -+ -+;; True is this is valid avr32 symbol operand -+(define_predicate "avr32_symbol_operand" -+ (ior (match_code "label_ref, symbol_ref") -+ (and (match_code "const") -+ (match_test "avr32_find_symbol(op)")))) -+ -+;; True is this is valid operand for the lda.w and call pseudo insns -+(define_predicate "avr32_address_operand" -+ (and (match_code "label_ref, symbol_ref") -+ (ior (match_test "TARGET_HAS_ASM_ADDR_PSEUDOS") -+ (match_test "flag_pic")) )) -+ -+;; True if this is a avr32 call operand -+(define_predicate "avr32_call_operand" -+ (ior (ior (match_operand 0 "register_operand") -+ (ior (match_operand 0 "avr32_const_pool_ref_operand") -+ (match_operand 0 "avr32_address_operand"))) -+ (match_test "SYMBOL_REF_RCALL_FUNCTION_P(op)"))) -+ -+;; Return true for operators performing ALU operations -+ -+(define_predicate "alu_operator" -+ (match_code "ior, xor, and, plus, minus, ashift, lshiftrt, ashiftrt")) -+ -+(define_predicate "avr32_add_shift_immediate_operand" -+ (and (match_operand 0 "immediate_operand") -+ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ku02\")"))) -+ -+(define_predicate "avr32_cond_register_immediate_operand" -+ (ior (match_operand 0 "register_operand") -+ (and (match_operand 0 "immediate_operand") -+ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ks08\")")))) -+ -+(define_predicate "avr32_cond_immediate_operand" -+ (and (match_operand 0 "immediate_operand") -+ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'I', \"Is08\")"))) -diff -Nrup gcc-4.2.1/gcc/config/avr32/simd.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/simd.md ---- gcc-4.2.1/gcc/config/avr32/simd.md 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/simd.md 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,145 @@ -+;; AVR32 machine description file for SIMD instructions. -+;; Copyright 2003-2006 Atmel Corporation. -+;; -+;; Written by Ronny Pedersen, Atmel Norway, -+;; -+;; This file is part of GCC. -+;; -+;; This program is free software; you can redistribute it and/or modify -+;; it under the terms of the GNU General Public License as published by -+;; the Free Software Foundation; either version 2 of the License, or -+;; (at your option) any later version. -+;; -+;; This program is distributed in the hope that it will be useful, -+;; but WITHOUT ANY WARRANTY; without even the implied warranty of -+;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+;; GNU General Public License for more details. -+;; -+;; You should have received a copy of the GNU General Public License -+;; along with this program; if not, write to the Free Software -+;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. -+ -+;; -*- Mode: Scheme -*- -+ -+ -+;; Vector modes -+(define_mode_macro VECM [V2HI V4QI]) -+(define_mode_attr size [(V2HI "h") (V4QI "b")]) -+ -+(define_insn "add3" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (plus:VECM (match_operand:VECM 1 "register_operand" "r") -+ (match_operand:VECM 2 "register_operand" "r")))] -+ "TARGET_SIMD" -+ "padd.\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+ -+(define_insn "sub3" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (minus:VECM (match_operand:VECM 1 "register_operand" "r") -+ (match_operand:VECM 2 "register_operand" "r")))] -+ "TARGET_SIMD" -+ "psub.\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+ -+(define_insn "abs2" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (abs:VECM (match_operand:VECM 1 "register_operand" "r")))] -+ "TARGET_SIMD" -+ "pabs.s\t%0, %1" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "ashl3" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (ashift:VECM (match_operand:VECM 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku04")))] -+ "TARGET_SIMD" -+ "plsl.\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "ashr3" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (ashiftrt:VECM (match_operand:VECM 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku04")))] -+ "TARGET_SIMD" -+ "pasr.\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "lshr3" -+ [(set (match_operand:VECM 0 "register_operand" "=r") -+ (lshiftrt:VECM (match_operand:VECM 1 "register_operand" "r") -+ (match_operand:SI 2 "immediate_operand" "Ku04")))] -+ "TARGET_SIMD" -+ "plsr.\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "smaxv2hi3" -+ [(set (match_operand:V2HI 0 "register_operand" "=r") -+ (smax:V2HI (match_operand:V2HI 1 "register_operand" "r") -+ (match_operand:V2HI 2 "register_operand" "r")))] -+ -+ "TARGET_SIMD" -+ "pmax.sh\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "sminv2hi3" -+ [(set (match_operand:V2HI 0 "register_operand" "=r") -+ (smin:V2HI (match_operand:V2HI 1 "register_operand" "r") -+ (match_operand:V2HI 2 "register_operand" "r")))] -+ -+ "TARGET_SIMD" -+ "pmin.sh\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "umaxv4qi3" -+ [(set (match_operand:V4QI 0 "register_operand" "=r") -+ (umax:V4QI (match_operand:V4QI 1 "register_operand" "r") -+ (match_operand:V4QI 2 "register_operand" "r")))] -+ -+ "TARGET_SIMD" -+ "pmax.ub\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "uminv4qi3" -+ [(set (match_operand:V4QI 0 "register_operand" "=r") -+ (umin:V4QI (match_operand:V4QI 1 "register_operand" "r") -+ (match_operand:V4QI 2 "register_operand" "r")))] -+ -+ "TARGET_SIMD" -+ "pmin.ub\t%0, %1, %2" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+ -+(define_insn "addsubv2hi" -+ [(set (match_operand:V2HI 0 "register_operand" "=r") -+ (vec_concat:V2HI -+ (plus:HI (match_operand:HI 1 "register_operand" "r") -+ (match_operand:HI 2 "register_operand" "r")) -+ (minus:HI (match_dup 1) (match_dup 2))))] -+ "TARGET_SIMD" -+ "paddsub.h\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -+ -+(define_insn "subaddv2hi" -+ [(set (match_operand:V2HI 0 "register_operand" "=r") -+ (vec_concat:V2HI -+ (minus:HI (match_operand:HI 1 "register_operand" "r") -+ (match_operand:HI 2 "register_operand" "r")) -+ (plus:HI (match_dup 1) (match_dup 2))))] -+ "TARGET_SIMD" -+ "psubadd.h\t%0, %1:b, %2:b" -+ [(set_attr "length" "4") -+ (set_attr "type" "alu")]) -diff -Nrup gcc-4.2.1/gcc/config/avr32/sync.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/sync.md ---- gcc-4.2.1/gcc/config/avr32/sync.md 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/sync.md 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,175 @@ -+;;================================================================= -+;; Atomic operations -+;;================================================================= -+ -+ -+(define_insn "sync_compare_and_swapsi" -+ [(set (match_operand:SI 0 "register_operand" "=&r,&r") -+ (match_operand:SI 1 "memory_operand" "+RKs16,+RKs16")) -+ (set (match_dup 1) -+ (unspec_volatile:SI -+ [(match_dup 1) -+ (match_operand:SI 2 "register_immediate_operand" "r,Ks21") -+ (match_operand:SI 3 "register_operand" "r,r")] -+ VUNSPEC_SYNC_CMPXCHG)) ] -+ "" -+ "0: -+ ssrf\t5 -+ ld.w\t%0,%1 -+ cp.w\t%0,%2 -+ brne\t0f -+ stcond\t%1, %3 -+ brne\t0b -+ 0: -+ " -+ [(set_attr "length" "16,18") -+ (set_attr "cc" "clobber")] -+ ) -+ -+ -+(define_code_macro atomic_op [plus minus and ior xor]) -+(define_code_attr atomic_asm_insn [(plus "add") (minus "sub") (and "and") (ior "or") (xor "eor")]) -+(define_code_attr atomic_insn [(plus "add") (minus "sub") (and "and") (ior "ior") (xor "xor")]) -+ -+(define_insn "sync_loadsi" -+ [(set (match_operand:SI 0 "register_operand" "=r") -+ (unspec_volatile:SI -+ [(match_operand:SI 1 "memory_operand" "RKs16") -+ (label_ref (match_operand 2 "" ""))] -+ VUNSPEC_SYNC_SET_LOCK_AND_LOAD) )] -+ "" -+ "%2: -+ ssrf\t5 -+ ld.w\t%0,%1" -+ [(set_attr "length" "6") -+ (set_attr "cc" "clobber")] -+ ) -+ -+(define_insn "sync_store_if_lock" -+ [(set (match_operand:SI 0 "memory_operand" "=RKs16") -+ (unspec_volatile:SI -+ [(match_operand:SI 1 "register_operand" "r") -+ (label_ref (match_operand 2 "" ""))] -+ VUNSPEC_SYNC_STORE_IF_LOCK) )] -+ "" -+ "stcond\t%0, %1 -+ brne\t%2" -+ [(set_attr "length" "6") -+ (set_attr "cc" "clobber")] -+ ) -+ -+ -+(define_expand "sync_si" -+ [(set (match_dup 2) -+ (unspec_volatile:SI -+ [(match_operand:SI 0 "memory_operand" "") -+ (match_dup 3)] -+ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) -+ (set (match_dup 2) -+ (atomic_op:SI (match_dup 2) -+ (match_operand:SI 1 "register_immediate_operand" ""))) -+ (set (match_dup 0) -+ (unspec_volatile:SI -+ [(match_dup 2) -+ (match_dup 3)] -+ VUNSPEC_SYNC_STORE_IF_LOCK) )] -+ "" -+ { -+ operands[2] = gen_reg_rtx (SImode); -+ operands[3] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); -+ } -+ ) -+ -+ -+ -+(define_expand "sync_old_si" -+ [(set (match_operand:SI 0 "register_operand" "") -+ (unspec_volatile:SI -+ [(match_operand:SI 1 "memory_operand" "") -+ (match_dup 4)] -+ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) -+ (set (match_dup 3) -+ (atomic_op:SI (match_dup 0) -+ (match_operand:SI 2 "register_immediate_operand" ""))) -+ (set (match_dup 1) -+ (unspec_volatile:SI -+ [(match_dup 3) -+ (match_dup 4)] -+ VUNSPEC_SYNC_STORE_IF_LOCK) )] -+ "" -+ { -+ operands[3] = gen_reg_rtx (SImode); -+ operands[4] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); -+ } -+ ) -+ -+(define_expand "sync_new_si" -+ [(set (match_operand:SI 0 "register_operand" "") -+ (unspec_volatile:SI -+ [(match_operand:SI 1 "memory_operand" "") -+ (match_dup 3)] -+ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) -+ (set (match_dup 0) -+ (atomic_op:SI (match_dup 0) -+ (match_operand:SI 2 "register_immediate_operand" ""))) -+ (set (match_dup 1) -+ (unspec_volatile:SI -+ [(match_dup 0) -+ (match_dup 3)] -+ VUNSPEC_SYNC_STORE_IF_LOCK) )] -+ "" -+ { -+ operands[3] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); -+ } -+ ) -+ -+ -+;(define_insn "sync_si" -+; [(set (match_operand:SI 0 "memory_operand" "+RKs16") -+; (unspec_volatile:SI -+; [(atomic_op:SI (match_dup 0) -+; (match_operand:SI 1 "register_operand" "r"))] -+; VUNSPEC_SYNC_CMPXCHG)) -+; (clobber (match_scratch:SI 2 "=&r"))] -+; "" -+; "0: -+; ssrf\t5 -+; ld.w\t%2,%0 -+; \t%2,%1 -+; stcond\t%0, %2 -+; brne\t0b -+; " -+; [(set_attr "length" "14") -+; (set_attr "cc" "clobber")] -+; ) -+; -+;(define_insn "sync_new_si" -+; [(set (match_operand:SI 1 "memory_operand" "+RKs16") -+; (unspec_volatile:SI -+; [(atomic_op:SI (match_dup 1) -+; (match_operand:SI 2 "register_operand" "r"))] -+; VUNSPEC_SYNC_CMPXCHG)) -+; (set (match_operand:SI 0 "register_operand" "=&r") -+; (atomic_op:SI (match_dup 1) -+; (match_dup 2)))] -+; "" -+; "0: -+; ssrf\t5 -+; ld.w\t%0,%1 -+; \t%0,%2 -+; stcond\t%1, %0 -+; brne\t0b -+; " -+; [(set_attr "length" "14") -+; (set_attr "cc" "clobber")] -+; ) -+ -+(define_insn "sync_lock_test_and_setsi" -+ [ (set (match_operand:SI 0 "register_operand" "=&r") -+ (match_operand:SI 1 "memory_operand" "+RKu00")) -+ (set (match_dup 1) -+ (match_operand:SI 2 "register_operand" "r")) ] -+ "" -+ "xchg\t%0, %p1, %2" -+ [(set_attr "length" "4")] -+ ) -diff -Nrup gcc-4.2.1/gcc/config/avr32/t-avr32 gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-avr32 ---- gcc-4.2.1/gcc/config/avr32/t-avr32 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-avr32 2007-09-28 10:33:00.000000000 +0200 -@@ -0,0 +1,76 @@ -+ -+MD_INCLUDES= $(srcdir)/config/avr32/avr32.md \ -+ $(srcdir)/config/avr32/sync.md \ -+ $(srcdir)/config/avr32/fpcp.md \ -+ $(srcdir)/config/avr32/simd.md \ -+ $(srcdir)/config/avr32/predicates.md -+ -+s-config s-conditions s-flags s-codes s-constants s-emit s-recog s-preds \ -+ s-opinit s-extract s-peep s-attr s-attrtab s-output: $(MD_INCLUDES) -+ -+# We want fine grained libraries, so use the new code -+# to build the floating point emulation libraries. -+FPBIT = fp-bit.c -+DPBIT = dp-bit.c -+ -+LIB1ASMSRC = avr32/lib1funcs.S -+LIB1ASMFUNCS = _avr32_f64_mul _avr32_f64_addsub _avr32_f64_addsub_fast _avr32_f64_to_u32 \ -+ _avr32_f64_to_s32 _avr32_f64_to_u64 _avr32_f64_to_s64 _avr32_u32_to_f64 \ -+ _avr32_s32_to_f64 _avr32_f64_cmp_eq _avr32_f64_cmp_ge _avr32_f64_cmp_lt \ -+ _avr32_f32_cmp_eq _avr32_f32_cmp_ge _avr32_f32_cmp_lt _avr32_f64_div \ -+ _avr32_f32_div _avr32_f32_div_fast _avr32_f32_addsub _avr32_f32_addsub_fast \ -+ _avr32_f32_mul _avr32_s32_to_f32 _avr32_u32_to_f32 _avr32_f32_to_s32 \ -+ _avr32_f32_to_u32 _avr32_f32_to_f64 _avr32_f64_to_f32 -+ -+#LIB2FUNCS_EXTRA += $(srcdir)/config/avr32/lib2funcs.S -+ -+MULTILIB_OPTIONS = march=ap/march=uc -+MULTILIB_DIRNAMES = ap uc -+MULTILIB_EXCEPTIONS = -+MULTILIB_MATCHES = march?ap=mcpu?ap7000 -+MULTILIB_MATCHES += march?ap=mcpu?ap7010 -+MULTILIB_MATCHES += march?ap=mcpu?ap7020 -+MULTILIB_MATCHES += march?uc=mcpu?uc3a0256 -+MULTILIB_MATCHES += march?uc=mcpu?uc3a0512 -+MULTILIB_MATCHES += march?uc=mcpu?uc3a1128 -+MULTILIB_MATCHES += march?uc=mcpu?uc3a1256 -+MULTILIB_MATCHES += march?uc=mcpu?uc3a1512 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b064 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b0128 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b0256 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b164 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b1128 -+MULTILIB_MATCHES += march?uc=mcpu?uc3b1256 -+MULTILIB_MATCHES += march?ap=mpart?ap7000 -+MULTILIB_MATCHES += march?ap=mpart?ap7010 -+MULTILIB_MATCHES += march?ap=mpart?ap7020 -+MULTILIB_MATCHES += march?uc=mpart?uc3a0256 -+MULTILIB_MATCHES += march?uc=mpart?uc3a0512 -+MULTILIB_MATCHES += march?uc=mpart?uc3a1128 -+MULTILIB_MATCHES += march?uc=mpart?uc3a1256 -+MULTILIB_MATCHES += march?uc=mpart?uc3a1512 -+MULTILIB_MATCHES += march?uc=mpart?uc3b064 -+MULTILIB_MATCHES += march?uc=mpart?uc3b0128 -+MULTILIB_MATCHES += march?uc=mpart?uc3b0256 -+MULTILIB_MATCHES += march?uc=mpart?uc3b164 -+MULTILIB_MATCHES += march?uc=mpart?uc3b1128 -+MULTILIB_MATCHES += march?uc=mpart?uc3b1256 -+ -+EXTRA_MULTILIB_PARTS = crtbegin.o crtbeginS.o crtend.o crtendS.o crti.o crtn.o -+ -+CRTSTUFF_T_CFLAGS = -mrelax -+CRTSTUFF_T_CFLAGS_S = -mrelax -fPIC -+TARGET_LIBGCC2_CFLAGS += -mrelax -+ -+LIBGCC = stmp-multilib -+INSTALL_LIBGCC = install-multilib -+ -+fp-bit.c: $(srcdir)/config/fp-bit.c -+ echo '#define FLOAT' > fp-bit.c -+ cat $(srcdir)/config/fp-bit.c >> fp-bit.c -+ -+dp-bit.c: $(srcdir)/config/fp-bit.c -+ cat $(srcdir)/config/fp-bit.c > dp-bit.c -+ -+ -+ -diff -Nrup gcc-4.2.1/gcc/config/avr32/t-elf gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-elf ---- gcc-4.2.1/gcc/config/avr32/t-elf 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-elf 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,16 @@ -+ -+# Assemble startup files. -+$(T)crti.o: $(srcdir)/config/avr32/crti.asm $(GCC_PASSES) -+ $(GCC_FOR_TARGET) $(CRTSTUFF_CFLAGS) $(CRTSTUFF_T_CFLAGS) $(INCLUDES) \ -+ -c -o $(T)crti.o -x assembler-with-cpp $(srcdir)/config/avr32/crti.asm -+ -+$(T)crtn.o: $(srcdir)/config/avr32/crtn.asm $(GCC_PASSES) -+ $(GCC_FOR_TARGET) $(CRTSTUFF_CFLAGS) $(CRTSTUFF_T_CFLAGS) $(INCLUDES) \ -+ -c -o $(T)crtn.o -x assembler-with-cpp $(srcdir)/config/avr32/crtn.asm -+ -+ -+# Build the libraries for both hard and soft floating point -+EXTRA_MULTILIB_PARTS = crtbegin.o crtbeginS.o crtend.o crtendS.o crti.o crtn.o -+ -+LIBGCC = stmp-multilib -+INSTALL_LIBGCC = install-multilib -diff -Nrup gcc-4.2.1/gcc/config/avr32/uclinux-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/uclinux-elf.h ---- gcc-4.2.1/gcc/config/avr32/uclinux-elf.h 1970-01-01 01:00:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/uclinux-elf.h 2007-05-07 14:29:10.000000000 +0200 -@@ -0,0 +1,20 @@ -+ -+/* Run-time Target Specification. */ -+#undef TARGET_VERSION -+#define TARGET_VERSION fputs (" (AVR32 uClinux with ELF)", stderr) -+ -+/* We don't want a .jcr section on uClinux. As if this makes a difference... */ -+#define TARGET_USE_JCR_SECTION 0 -+ -+/* Here we go. Drop the crtbegin/crtend stuff completely. */ -+#undef STARTFILE_SPEC -+#define STARTFILE_SPEC \ -+ "%{!shared: %{pg:gcrt1.o%s} %{!pg:%{p:gcrt1.o%s}" \ -+ " %{!p:%{profile:gcrt1.o%s}" \ -+ " %{!profile:crt1.o%s}}}} crti.o%s" -+ -+#undef ENDFILE_SPEC -+#define ENDFILE_SPEC "crtn.o%s" -+ -+#undef TARGET_DEFAULT -+#define TARGET_DEFAULT (AVR32_FLAG_NO_INIT_GOT) -diff -Nrup gcc-4.2.1/gcc/config/host-linux.c gcc-4.2.1.atmel.1.3.2/gcc/config/host-linux.c ---- gcc-4.2.1/gcc/config/host-linux.c 2005-08-01 19:43:33.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config/host-linux.c 2007-05-07 14:29:15.000000000 +0200 -@@ -26,6 +26,9 @@ - #include "hosthooks.h" - #include "hosthooks-def.h" - -+#ifndef SSIZE_MAX -+#define SSIZE_MAX LONG_MAX -+#endif - - /* Linux has a feature called exec-shield-randomize that perturbs the - address of non-fixed mapped segments by a (relatively) small amount. -diff -Nrup gcc-4.2.1/gcc/config.gcc gcc-4.2.1.atmel.1.3.2/gcc/config.gcc ---- gcc-4.2.1/gcc/config.gcc 2007-02-03 06:25:20.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/config.gcc 2007-09-28 10:33:09.000000000 +0200 -@@ -782,6 +782,24 @@ avr-*-*) - tm_file="avr/avr.h dbxelf.h" - use_fixproto=yes - ;; -+avr32*-*-linux*) -+ tm_file="dbxelf.h elfos.h linux.h avr32/linux-elf.h avr32/avr32.h " -+ tmake_file="t-linux avr32/t-avr32 avr32/t-elf" -+ extra_parts="crtbegin.o crtbeginS.o crtend.o crtendS.o" -+ extra_modes=avr32/avr32-modes.def -+ gnu_ld=yes -+ ;; -+avr32*-*-uclinux*) -+ tm_file="dbxelf.h elfos.h linux.h avr32/linux-elf.h avr32/uclinux-elf.h avr32/avr32.h" -+ tmake_file="t-linux avr32/t-avr32 avr32/t-elf" -+ extra_modes=avr32/avr32-modes.def -+ gnu_ld=yes -+ ;; -+avr32-*-*) -+ tm_file="dbxelf.h elfos.h avr32/avr32.h avr32/avr32-elf.h" -+ tmake_file="avr32/t-avr32 avr32/t-elf" -+ extra_modes=avr32/avr32-modes.def -+ ;; - bfin*-elf*) - tm_file="${tm_file} dbxelf.h elfos.h bfin/elf.h" - tmake_file=bfin/t-bfin-elf -@@ -1682,6 +1700,9 @@ pdp11-*-bsd) - pdp11-*-*) - use_fixproto=yes - ;; -+avr-*-*) -+ use_fixproto=yes -+ ;; - # port not yet contributed - #powerpc-*-openbsd*) - # tmake_file="${tmake_file} rs6000/t-fprules rs6000/t-fprules-fpbit " -@@ -2718,6 +2739,32 @@ case "${target}" in - fi - ;; - -+ avr32*-*-*) -+ supported_defaults="part arch" -+ -+ case "$with_part" in -+ "" \ -+ | "ap7000" | "ap7010" | "ap7020" | "uc3a0256" | "uc3a0512" | "uc3a1128" | "uc3a1256" | "uc3a1512" ) -+ # OK -+ ;; -+ *) -+ echo "Unknown part used in --with-part=$with_part" 1>&2 -+ exit 1 -+ ;; -+ esac -+ -+ case "$with_arch" in -+ "" \ -+ | "ap" | "uc") -+ # OK -+ ;; -+ *) -+ echo "Unknown arch used in --with-arch=$with_arch" 1>&2 -+ exit 1 -+ ;; -+ esac -+ ;; -+ - fr*-*-*linux*) - supported_defaults=cpu - case "$with_cpu" in -diff -Nrup gcc-4.2.1/gcc/doc/extend.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/extend.texi ---- gcc-4.2.1/gcc/doc/extend.texi 2007-03-12 23:10:12.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/doc/extend.texi 2007-09-28 10:32:30.000000000 +0200 -@@ -1981,7 +1981,7 @@ this attribute to work correctly. - - @item interrupt - @cindex interrupt handler functions --Use this attribute on the ARM, AVR, C4x, CRX, M32C, M32R/D, MS1, and Xstormy16 -+Use this attribute on the ARM, AVR, AVR32, C4x, CRX, M32C, M32R/D, MS1, and Xstormy16 - ports to indicate that the specified function is an interrupt handler. - The compiler will generate function entry and exit sequences suitable - for use in an interrupt handler when this attribute is present. -@@ -2000,6 +2000,15 @@ void f () __attribute__ ((interrupt ("IR - - Permissible values for this parameter are: IRQ, FIQ, SWI, ABORT and UNDEF@. - -+Note, for the AVR32, you can specify which banking scheme is used for -+the interrupt mode this interrupt handler is used in like this: -+ -+@smallexample -+void f () __attribute__ ((interrupt ("FULL"))); -+@end smallexample -+ -+Permissible values for this parameter are: FULL, HALF, NONE and UNDEF. -+ - @item interrupt_handler - @cindex interrupt handler functions on the Blackfin, m68k, H8/300 and SH processors - Use this attribute on the Blackfin, m68k, H8/300, H8/300H, H8S, and SH to -@@ -6167,6 +6176,7 @@ instructions, but allow the compiler to - @menu - * Alpha Built-in Functions:: - * ARM Built-in Functions:: -+* AVR32 Built-in Functions:: - * Blackfin Built-in Functions:: - * FR-V Built-in Functions:: - * X86 Built-in Functions:: -@@ -6405,6 +6415,54 @@ long long __builtin_arm_wxor (long long, - long long __builtin_arm_wzero () - @end smallexample - -+@node AVR32 Built-in Functions -+@subsection AVR32 Built-in Functions -+ -+ -+@smallexample -+ -+int __builtin_sats (int /*Rd*/,int /*sa*/, int /*bn*/) -+int __builtin_satu (int /*Rd*/,int /*sa*/, int /*bn*/) -+int __builtin_satrnds (int /*Rd*/,int /*sa*/, int /*bn*/) -+int __builtin_satrndu (int /*Rd*/,int /*sa*/, int /*bn*/) -+short __builtin_mulsathh_h (short, short) -+int __builtin_mulsathh_w (short, short) -+short __builtin_mulsatrndhh_h (short, short) -+int __builtin_mulsatrndwh_w (int, short) -+int __builtin_mulsatwh_w (int, short) -+int __builtin_macsathh_w (int, short, short) -+short __builtin_satadd_h (short, short) -+short __builtin_satsub_h (short, short) -+int __builtin_satadd_w (int, int) -+int __builtin_satsub_w (int, int) -+long long __builtin_mulwh_d(int, short) -+long long __builtin_mulnwh_d(int, short) -+long long __builtin_macwh_d(long long, int, short) -+long long __builtin_machh_d(long long, short, short) -+ -+void __builtin_musfr(int); -+int __builtin_mustr(void); -+int __builtin_mfsr(int /*Status Register Address*/) -+void __builtin_mtsr(int /*Status Register Address*/, int /*Value*/) -+int __builtin_mfdr(int /*Debug Register Address*/) -+void __builtin_mtdr(int /*Debug Register Address*/, int /*Value*/) -+void __builtin_cache(void * /*Address*/, int /*Cache Operation*/) -+void __builtin_sync(int /*Sync Operation*/) -+void __builtin_tlbr(void) -+void __builtin_tlbs(void) -+void __builtin_tlbw(void) -+void __builtin_breakpoint(void) -+int __builtin_xchg(void * /*Address*/, int /*Value*/ ) -+short __builtin_bswap_16(short) -+int __builtin_bswap_32(int) -+void __builtin_cop(int/*cpnr*/, int/*crd*/, int/*crx*/, int/*cry*/, int/*op*/) -+int __builtin_mvcr_w(int/*cpnr*/, int/*crs*/) -+void __builtin_mvrc_w(int/*cpnr*/, int/*crd*/, int/*value*/) -+long long __builtin_mvcr_d(int/*cpnr*/, int/*crs*/) -+void __builtin_mvrc_d(int/*cpnr*/, int/*crd*/, long long/*value*/) -+ -+@end smallexample -+ - @node Blackfin Built-in Functions - @subsection Blackfin Built-in Functions - -diff -Nrup gcc-4.2.1/gcc/doc/invoke.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/invoke.texi ---- gcc-4.2.1/gcc/doc/invoke.texi 2007-04-24 23:54:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/doc/invoke.texi 2007-09-28 10:32:30.000000000 +0200 -@@ -190,7 +190,7 @@ in the following sections. - -fno-default-inline -fvisibility-inlines-hidden @gol - -Wabi -Wctor-dtor-privacy @gol - -Wnon-virtual-dtor -Wreorder @gol ---Weffc++ -Wno-deprecated -Wstrict-null-sentinel @gol -+-Weffc++ -Wno-deprecated @gol - -Wno-non-template-friend -Wold-style-cast @gol - -Woverloaded-virtual -Wno-pmf-conversions @gol - -Wsign-promo} -@@ -588,6 +588,12 @@ Objective-C and Objective-C++ Dialects}. - -mauto-incdec -minmax -mlong-calls -mshort @gol - -msoft-reg-count=@var{count}} - -+@emph{AVR32 Options} -+@gccoptlist{-muse-rodata-section -mhard-float -msoft-float -mrelax @gol -+-mforce-double-align -mno-init-got -mpart=@var{part} -mcpu=@var{cpu} @gol -+-march=@var{arch} -mfast-float -masm-addr-pseudos -mno-asm-addr-pseudos -mno-pic -+} -+ - @emph{MCore Options} - @gccoptlist{-mhardlit -mno-hardlit -mdiv -mno-div -mrelax-immediates @gol - -mno-relax-immediates -mwide-bitfields -mno-wide-bitfields @gol -@@ -1868,14 +1874,6 @@ to filter out those warnings. - @opindex Wno-deprecated - Do not warn about usage of deprecated features. @xref{Deprecated Features}. - --@item -Wstrict-null-sentinel @r{(C++ only)} --@opindex Wstrict-null-sentinel --Warn also about the use of an uncasted @code{NULL} as sentinel. When --compiling only with GCC this is a valid sentinel, as @code{NULL} is defined --to @code{__null}. Although it is a null pointer constant not a null pointer, --it is guaranteed to of the same size as a pointer. But this use is --not portable across different compilers. -- - @item -Wno-non-template-friend @r{(C++ only)} - @opindex Wno-non-template-friend - Disable warnings when non-templatized friend functions are declared -@@ -2732,13 +2730,11 @@ requiring @option{-O}. - If you want to warn about code which uses the uninitialized value of the - variable in its own initializer, use the @option{-Winit-self} option. - --These warnings occur for individual uninitialized or clobbered --elements of structure, union or array variables as well as for --variables which are uninitialized or clobbered as a whole. They do --not occur for variables or elements declared @code{volatile}. Because --these warnings depend on optimization, the exact variables or elements --for which there are warnings will depend on the precise optimization --options and version of GCC used. -+These warnings occur only for variables that are candidates for -+register allocation. Therefore, they do not occur for a variable that -+is declared @code{volatile}, or whose address is taken, or whose size -+is other than 1, 2, 4 or 8 bytes. Also, they do not occur for -+structures, unions or arrays, even when they are in registers. - - Note that there may be no warning about a variable that is used only - to compute a value that itself is never used, because such -@@ -6201,10 +6197,6 @@ If number of candidates in the set is sm - we always try to remove unnecessary ivs from the set during its - optimization when a new iv is added to the set. - --@item scev-max-expr-size --Bound on size of expressions used in the scalar evolutions analyzer. --Large expressions slow the analyzer. -- - @item vect-max-version-checks - The maximum number of runtime checks that can be performed when doing - loop versioning in the vectorizer. See option ftree-vect-loop-version -@@ -7402,7 +7394,7 @@ platform. - * ARC Options:: - * ARM Options:: - * AVR Options:: --* Blackfin Options:: -+* AVR32 Options:: - * CRIS Options:: - * CRX Options:: - * Darwin Options:: -@@ -7867,81 +7859,68 @@ comply to the C standards, but it will p - size. - @end table - --@node Blackfin Options --@subsection Blackfin Options --@cindex Blackfin Options -+@node AVR32 Options -+@subsection AVR32 Options -+@cindex AVR32 Options -+ -+These options are defined for AVR32 implementations: - - @table @gcctabopt --@item -momit-leaf-frame-pointer --@opindex momit-leaf-frame-pointer --Don't keep the frame pointer in a register for leaf functions. This --avoids the instructions to save, set up and restore frame pointers and --makes an extra register available in leaf functions. The option --@option{-fomit-frame-pointer} removes the frame pointer for all functions --which might make debugging harder. -+@item -muse-rodata-section -+@opindex muse-rodata-section -+Use section @samp{.rodata} for read-only data instead of @samp{.text}. - --@item -mspecld-anomaly --@opindex mspecld-anomaly --When enabled, the compiler will ensure that the generated code does not --contain speculative loads after jump instructions. This option is enabled --by default. -- --@item -mno-specld-anomaly --@opindex mno-specld-anomaly --Don't generate extra code to prevent speculative loads from occurring. -- --@item -mcsync-anomaly --@opindex mcsync-anomaly --When enabled, the compiler will ensure that the generated code does not --contain CSYNC or SSYNC instructions too soon after conditional branches. --This option is enabled by default. -- --@item -mno-csync-anomaly --@opindex mno-csync-anomaly --Don't generate extra code to prevent CSYNC or SSYNC instructions from --occurring too soon after a conditional branch. -- --@item -mlow-64k --@opindex mlow-64k --When enabled, the compiler is free to take advantage of the knowledge that --the entire program fits into the low 64k of memory. -- --@item -mno-low-64k --@opindex mno-low-64k --Assume that the program is arbitrarily large. This is the default. -+@item -mrelax -+@opindex mrelax -+Enable relaxing in linker. This means that when the address of symbols -+are known at link time, the linker can optimize @samp{icall} and @samp{mcall} -+instructions into a @samp{rcall} instruction if possible. Loading the address -+of a symbol can also be optimized. -+ -+@item -mforce-double-align -+@opindex mforce-double-align -+Force double-word alignment for double-word memory accesses. -+ -+@item -mno-init-got -+@opindex mno-init-got -+Do not initialize the GOT register before using it when compiling PIC -+code. - --@item -mid-shared-library --@opindex mid-shared-library --Generate code that supports shared libraries via the library ID method. --This allows for execute in place and shared libraries in an environment --without virtual memory management. This option implies @option{-fPIC}. -+@item -mno-pic -+@opindex mno-pic -+Do not emit position-independent code (will break dynamic linking.) - --@item -mno-id-shared-library --@opindex mno-id-shared-library --Generate code that doesn't assume ID based shared libraries are being used. --This is the default. -+@item -masm-addr-pseudos -+@opindex masm-addr-pseudos -+Use assembler pseudo-instructions lda.w and call for handling direct -+addresses. (Enabled by default) -+ -+@item -mno-asm-addr-pseudos -+@opindex mno-asm-addr-pseudos -+Do not use assembler pseudo-instructions lda.w and call for handling direct addresses. -+ -+@item -mpart=@var{part} -+@opindex mpart -+Generate code for the specified part. Permissible parts are: @samp{ap7000}, -+@samp{ap7010},@samp{ap7020},@samp{uc3a0256}, @samp{uc3a0512}, -+@samp{uc3a1128}, @samp{uc3a1256}, @samp{uc3a1512}. - --@item -mshared-library-id=n --@opindex mshared-library-id --Specified the identification number of the ID based shared library being --compiled. Specifying a value of 0 will generate more compact code, specifying --other values will force the allocation of that number to the current --library but is no more space or time efficient than omitting this option. -+@item -mcpu=@var{cpu-type} -+@opindex mcpu -+Same as -mpart. Obsolete. -+ -+@item -march=@var{arch} -+@opindex march -+Generate code for the specified architecture. Permissible architectures are: -+@samp{ap} and @samp{uc}. -+ -+@item -mfast-float -+@opindex mfast-float -+Enable fast floating-point library that does not conform to ieee but is still good enough -+for most applications. The fast floating-point library does not round to the nearest even -+but away from zero. Enabled by default if the -funsafe-math-optimizations switch is specified. - --@item -mlong-calls --@itemx -mno-long-calls --@opindex mlong-calls --@opindex mno-long-calls --Tells the compiler to perform function calls by first loading the --address of the function into a register and then performing a subroutine --call on this register. This switch is needed if the target function --will lie outside of the 24 bit addressing range of the offset based --version of subroutine call instruction. - --This feature is not enabled by default. Specifying --@option{-mno-long-calls} will restore the default behavior. Note these --switches have no effect on how the compiler generates code to handle --function calls via function pointers. - @end table - - @node CRIS Options -@@ -11852,6 +11831,7 @@ conventions that adheres to the March 19 - Application Binary Interface, PowerPC processor supplement. This is the - default unless you configured GCC using @samp{powerpc-*-eabiaix}. - -+ - @item -mcall-sysv-eabi - @opindex mcall-sysv-eabi - Specify both @option{-mcall-sysv} and @option{-meabi} options. -@@ -14153,4 +14133,4 @@ You need to include the special files al - exist, because otherwise they won't get converted. - - @xref{Protoize Caveats}, for more information on how to use --@code{protoize} successfully. -+@code{protoize} successfully. -\ No newline at end of file -diff -Nrup gcc-4.2.1/gcc/doc/md.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/md.texi ---- gcc-4.2.1/gcc/doc/md.texi 2007-04-04 03:24:10.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/doc/md.texi 2007-09-28 10:32:30.000000000 +0200 -@@ -3,6 +3,7 @@ - @c This is part of the GCC manual. - @c For copying conditions, see the file gcc.texi. - -+ - @ifset INTERNALS - @node Machine Desc - @chapter Machine Descriptions -@@ -1681,6 +1682,58 @@ A memory reference suitable for iWMMXt l - A memory reference suitable for the ARMv4 ldrsb instruction. - @end table - -+@item AVR32 family---@file{avr32.h} -+@table @code -+@item f -+Floating-point registers (f0 to f15) -+ -+@item Ku@var{bits} -+Unsigned constant representable with @var{bits} number of bits (Must be -+two digits). I.e: An unsigned 8-bit constant is written as @samp{Ku08} -+ -+@item Ks@var{bits} -+Signed constant representable with @var{bits} number of bits (Must be -+two digits). I.e: A signed 12-bit constant is written as @samp{Ks12} -+ -+@item Is@var{bits} -+The negated range of a signed constant representable with @var{bits} -+number of bits. The same as @samp{Ks@var{bits}} with a negated range. -+This means that the constant must be in the range @math{-2^{bits-1}-1} to @math{2^{bits-1}} -+ -+@item G -+A single/double precision floating-point immediate or 64-bit integer -+immediate where the least and most significant words both can be -+loaded with a move instruction. That is the the integer form of the -+values in the least and most significant words both are in the range -+@math{-2^{20}} to @math{2^{20}-1}. -+ -+@item RKs@var{bits} -+A memory reference where the address consists of a base register -+plus a signed immediate displacement with range given by @samp{Ks@var{bits}} -+which has the same format as for the signed immediate integer constraint -+given above. -+ -+@item RKu@var{bits} -+A memory reference where the address consists of a base register -+plus an unsigned immediate displacement with range given by @samp{Ku@var{bits}} -+which has the same format as for the unsigned immediate integer constraint -+given above. -+ -+@item S -+A memory reference with an immediate or register offset -+ -+@item T -+A memory reference to a constant pool entry -+ -+@item W -+A valid operand for use in the @samp{lda.w} instruction macro when -+relaxing is enabled -+ -+@item Z -+A memory reference valid for coprocessor memory instructions -+ -+@end table -+ - @item AVR family---@file{config/avr/constraints.md} - @table @code - @item l -diff -Nrup gcc-4.2.1/gcc/expmed.c gcc-4.2.1.atmel.1.3.2/gcc/expmed.c ---- gcc-4.2.1/gcc/expmed.c 2007-04-25 07:43:32.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/expmed.c 2007-09-28 10:33:09.000000000 +0200 -@@ -37,6 +37,7 @@ Software Foundation, 51 Franklin Street, - #include "real.h" - #include "recog.h" - #include "langhooks.h" -+#include "target.h" - - static void store_fixed_bit_field (rtx, unsigned HOST_WIDE_INT, - unsigned HOST_WIDE_INT, -@@ -455,9 +456,18 @@ store_bit_field (rtx str_rtx, unsigned H - ? ((GET_MODE_SIZE (fieldmode) >= UNITS_PER_WORD - || GET_MODE_SIZE (GET_MODE (op0)) == GET_MODE_SIZE (fieldmode)) - && byte_offset % GET_MODE_SIZE (fieldmode) == 0) -- : (! SLOW_UNALIGNED_ACCESS (fieldmode, MEM_ALIGN (op0)) -- || (offset * BITS_PER_UNIT % bitsize == 0 -- && MEM_ALIGN (op0) % GET_MODE_BITSIZE (fieldmode) == 0)))) -+ : ( -+ -+ /* NB! Added for AVR32, and I think this should be true for -+ all targets not using narrow volatile bitfields. If the -+ bitfield is volatile then we need to perform an access -+ consistent with the container type. */ -+ !(MEM_VOLATILE_P (op0) -+ && GET_MODE_BITSIZE (GET_MODE (op0)) != bitsize -+ && !targetm.narrow_volatile_bitfield ()) -+ && (! SLOW_UNALIGNED_ACCESS (fieldmode, MEM_ALIGN (op0)) -+ || (offset * BITS_PER_UNIT % bitsize == 0 -+ && MEM_ALIGN (op0) % GET_MODE_BITSIZE (fieldmode) == 0))))) - { - if (MEM_P (op0)) - op0 = adjust_address (op0, fieldmode, offset); -@@ -1257,6 +1267,13 @@ extract_bit_field (rtx str_rtx, unsigned - && GET_MODE_SIZE (mode1) != 0 - && byte_offset % GET_MODE_SIZE (mode1) == 0) - || (MEM_P (op0) -+ /* NB! Added for AVR32, and I think this should be true for -+ all targets not using narrow volatile bitfields. If the -+ bitfield is volatile then we need to perform an access -+ consistent with the container type. */ -+ && !(MEM_VOLATILE_P (op0) -+ && GET_MODE_BITSIZE (GET_MODE (op0)) != bitsize -+ && !targetm.narrow_volatile_bitfield ()) - && (! SLOW_UNALIGNED_ACCESS (mode, MEM_ALIGN (op0)) - || (offset * BITS_PER_UNIT % bitsize == 0 - && MEM_ALIGN (op0) % bitsize == 0))))) -diff -Nrup gcc-4.2.1/gcc/expr.c gcc-4.2.1.atmel.1.3.2/gcc/expr.c ---- gcc-4.2.1/gcc/expr.c 2007-04-23 03:57:48.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/expr.c 2007-09-28 10:33:08.000000000 +0200 -@@ -3520,18 +3520,19 @@ emit_single_push_insn (enum machine_mode - } - else - { -+ emit_move_insn (stack_pointer_rtx, -+ expand_binop (Pmode, - #ifdef STACK_GROWS_DOWNWARD -- /* ??? This seems wrong if STACK_PUSH_CODE == POST_DEC. */ -- dest_addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, -- GEN_INT (-(HOST_WIDE_INT) rounded_size)); -+ sub_optab, - #else -- /* ??? This seems wrong if STACK_PUSH_CODE == POST_INC. */ -- dest_addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, -- GEN_INT (rounded_size)); -+ add_optab, - #endif -- dest_addr = gen_rtx_PRE_MODIFY (Pmode, stack_pointer_rtx, dest_addr); -+ stack_pointer_rtx, -+ GEN_INT (rounded_size), -+ NULL_RTX, 0, OPTAB_LIB_WIDEN)); -+ dest_addr = stack_pointer_rtx; - } -- -+ - dest = gen_rtx_MEM (mode, dest_addr); - - if (type != 0) -@@ -5510,7 +5511,16 @@ store_field (rtx target, HOST_WIDE_INT b - is a bit field, we cannot use addressing to access it. - Use bit-field techniques or SUBREG to store in it. */ - -- if (mode == VOIDmode -+ if ( -+ /* NB! Added for AVR32, and I think this should be true for -+ all targets not using narrow volatile bitfields. If the -+ bitfield is volatile then we need to perform an access -+ consistent with the container type. */ -+ (MEM_P (target) -+ && MEM_VOLATILE_P (target) -+ && GET_MODE_BITSIZE (GET_MODE (target)) > bitsize -+ && !targetm.narrow_volatile_bitfield ()) -+ || mode == VOIDmode - || (mode != BLKmode && ! direct_store[(int) mode] - && GET_MODE_CLASS (mode) != MODE_COMPLEX_INT - && GET_MODE_CLASS (mode) != MODE_COMPLEX_FLOAT) -@@ -7512,7 +7522,16 @@ expand_expr_real_1 (tree exp, rtx target - by doing the extract into an object as wide as the field - (which we know to be the width of a basic mode), then - storing into memory, and changing the mode to BLKmode. */ -- if (mode1 == VOIDmode -+ if ( -+ /* NB! Added for AVR32, and I think this should be true for -+ all targets not using narrow volatile bitfields. If the -+ bitfield is volatile then we need to perform an access -+ consistent with the container type. */ -+ (MEM_P (op0) -+ && MEM_VOLATILE_P (op0) -+ && GET_MODE_BITSIZE (GET_MODE (op0)) > bitsize -+ && !targetm.narrow_volatile_bitfield ()) -+ || mode1 == VOIDmode - || REG_P (op0) || GET_CODE (op0) == SUBREG - || (mode1 != BLKmode && ! direct_load[(int) mode1] - && GET_MODE_CLASS (mode) != MODE_COMPLEX_INT -diff -Nrup gcc-4.2.1/gcc/function.c gcc-4.2.1.atmel.1.3.2/gcc/function.c ---- gcc-4.2.1/gcc/function.c 2007-07-19 05:25:32.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/function.c 2007-09-28 10:33:09.000000000 +0200 -@@ -2677,8 +2677,12 @@ assign_parm_setup_reg (struct assign_par - SET_DECL_RTL (parm, parmreg); - - /* Copy the value into the register. */ -- if (data->nominal_mode != data->passed_mode -- || promoted_nominal_mode != data->promoted_mode) -+ if ( (data->nominal_mode != data->passed_mode -+ /* Added for AVR32: If passed_mode is equal -+ to promoted nominal mode why should be convert? -+ The conversion should make no difference. */ -+ && data->passed_mode != promoted_nominal_mode) -+ || promoted_nominal_mode != data->promoted_mode) - { - int save_tree_used; - -diff -Nrup gcc-4.2.1/gcc/genemit.c gcc-4.2.1.atmel.1.3.2/gcc/genemit.c ---- gcc-4.2.1/gcc/genemit.c 2006-03-29 23:07:12.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/genemit.c 2007-09-28 10:33:08.000000000 +0200 -@@ -122,6 +122,24 @@ max_operand_vec (rtx insn, int arg) - } - - static void -+gen_vararg_prologue(int operands) -+{ -+ int i; -+ -+ if (operands > 1) -+ { -+ for (i = 1; i < operands; i++) -+ printf(" rtx operand%d ATTRIBUTE_UNUSED;\n", i); -+ -+ printf(" va_list args;\n\n"); -+ printf(" va_start(args, operand0);\n"); -+ for (i = 1; i < operands; i++) -+ printf(" operand%d = va_arg(args, rtx);\n", i); -+ printf(" va_end(args);\n\n"); -+ } -+} -+ -+static void - print_code (RTX_CODE code) - { - const char *p1; -@@ -406,18 +424,16 @@ gen_insn (rtx insn, int lineno) - fatal ("match_dup operand number has no match_operand"); - - /* Output the function name and argument declarations. */ -- printf ("rtx\ngen_%s (", XSTR (insn, 0)); -+ printf ("rtx\ngen_%s ", XSTR (insn, 0)); -+ - if (operands) -- for (i = 0; i < operands; i++) -- if (i) -- printf (",\n\trtx operand%d ATTRIBUTE_UNUSED", i); -- else -- printf ("rtx operand%d ATTRIBUTE_UNUSED", i); -+ printf("(rtx operand0 ATTRIBUTE_UNUSED, ...)\n"); - else -- printf ("void"); -- printf (")\n"); -+ printf("(void)\n"); - printf ("{\n"); - -+ gen_vararg_prologue(operands); -+ - /* Output code to construct and return the rtl for the instruction body. */ - - if (XVECLEN (insn, 1) == 1) -@@ -457,16 +473,12 @@ gen_expand (rtx expand) - operands = max_operand_vec (expand, 1); - - /* Output the function name and argument declarations. */ -- printf ("rtx\ngen_%s (", XSTR (expand, 0)); -+ printf ("rtx\ngen_%s ", XSTR (expand, 0)); - if (operands) -- for (i = 0; i < operands; i++) -- if (i) -- printf (",\n\trtx operand%d", i); -- else -- printf ("rtx operand%d", i); -+ printf("(rtx operand0 ATTRIBUTE_UNUSED, ...)\n"); - else -- printf ("void"); -- printf (")\n"); -+ printf("(void)\n"); -+ - printf ("{\n"); - - /* If we don't have any C code to write, only one insn is being written, -@@ -476,6 +488,8 @@ gen_expand (rtx expand) - && operands > max_dup_opno - && XVECLEN (expand, 1) == 1) - { -+ gen_vararg_prologue(operands); -+ - printf (" return "); - gen_exp (XVECEXP (expand, 1, 0), DEFINE_EXPAND, NULL); - printf (";\n}\n\n"); -@@ -489,6 +503,7 @@ gen_expand (rtx expand) - for (; i <= max_scratch_opno; i++) - printf (" rtx operand%d ATTRIBUTE_UNUSED;\n", i); - printf (" rtx _val = 0;\n"); -+ gen_vararg_prologue(operands); - printf (" start_sequence ();\n"); - - /* The fourth operand of DEFINE_EXPAND is some code to be executed -diff -Nrup gcc-4.2.1/gcc/genflags.c gcc-4.2.1.atmel.1.3.2/gcc/genflags.c ---- gcc-4.2.1/gcc/genflags.c 2006-01-23 16:15:12.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/genflags.c 2007-09-28 10:33:09.000000000 +0200 -@@ -128,7 +128,6 @@ static void - gen_proto (rtx insn) - { - int num = num_operands (insn); -- int i; - const char *name = XSTR (insn, 0); - int truth = maybe_eval_c_test (XSTR (insn, 2)); - -@@ -159,12 +158,7 @@ gen_proto (rtx insn) - if (num == 0) - fputs ("void", stdout); - else -- { -- for (i = 1; i < num; i++) -- fputs ("rtx, ", stdout); -- -- fputs ("rtx", stdout); -- } -+ fputs("rtx, ...", stdout); - - puts (");"); - -@@ -174,12 +168,7 @@ gen_proto (rtx insn) - { - printf ("static inline rtx\ngen_%s", name); - if (num > 0) -- { -- putchar ('('); -- for (i = 0; i < num-1; i++) -- printf ("rtx ARG_UNUSED (%c), ", 'a' + i); -- printf ("rtx ARG_UNUSED (%c))\n", 'a' + i); -- } -+ puts("(rtx ARG_UNUSED(a), ...)"); - else - puts ("(void)"); - puts ("{\n return 0;\n}"); -diff -Nrup gcc-4.2.1/gcc/genoutput.c gcc-4.2.1.atmel.1.3.2/gcc/genoutput.c ---- gcc-4.2.1/gcc/genoutput.c 2006-03-29 23:07:12.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/genoutput.c 2007-09-28 10:33:08.000000000 +0200 -@@ -387,7 +387,7 @@ output_insn_data (void) - } - - if (d->name && d->name[0] != '*') -- printf (" (insn_gen_fn) gen_%s,\n", d->name); -+ printf (" gen_%s,\n", d->name); - else - printf (" 0,\n"); - -diff -Nrup gcc-4.2.1/gcc/ifcvt.c gcc-4.2.1.atmel.1.3.2/gcc/ifcvt.c ---- gcc-4.2.1/gcc/ifcvt.c 2006-11-15 09:37:38.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/ifcvt.c 2007-09-28 10:33:08.000000000 +0200 -@@ -1051,7 +1051,11 @@ noce_try_addcc (struct noce_if_info *if_ - != UNKNOWN)) - { - rtx cond = if_info->cond; -- enum rtx_code code = reversed_comparison_code (cond, if_info->jump); -+ /* This generates wrong code for AVR32. The cond code need not be reversed -+ since the addmodecc patterns add if the condition is NOT met. */ -+ /* enum rtx_code code = reversed_comparison_code (cond, if_info->jump);*/ -+ enum rtx_code code = GET_CODE(cond); -+ - - /* First try to use addcc pattern. */ - if (general_operand (XEXP (cond, 0), VOIDmode) -diff -Nrup gcc-4.2.1/gcc/longlong.h gcc-4.2.1.atmel.1.3.2/gcc/longlong.h ---- gcc-4.2.1/gcc/longlong.h 2006-06-13 19:44:56.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/longlong.h 2007-09-28 10:33:08.000000000 +0200 -@@ -227,6 +227,39 @@ UDItype __umulsidi3 (USItype, USItype); - #define UDIV_TIME 100 - #endif /* __arm__ */ - -+#if defined (__avr32__) && W_TYPE_SIZE == 32 -+#define add_ssaaaa(sh, sl, ah, al, bh, bl) \ -+ __asm__ ("add\t%1, %4, %5\n\tadc\t%0, %2, %3" \ -+ : "=r" ((USItype) (sh)), \ -+ "=&r" ((USItype) (sl)) \ -+ : "r" ((USItype) (ah)), \ -+ "r" ((USItype) (bh)), \ -+ "r" ((USItype) (al)), \ -+ "r" ((USItype) (bl)) __CLOBBER_CC) -+#define sub_ddmmss(sh, sl, ah, al, bh, bl) \ -+ __asm__ ("sub\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ -+ : "=r" ((USItype) (sh)), \ -+ "=&r" ((USItype) (sl)) \ -+ : "r" ((USItype) (ah)), \ -+ "r" ((USItype) (bh)), \ -+ "r" ((USItype) (al)), \ -+ "r" ((USItype) (bl)) __CLOBBER_CC) -+ -+#define __umulsidi3(a,b) ((UDItype)(a) * (UDItype)(b)) -+ -+#define umul_ppmm(w1, w0, u, v) \ -+{ \ -+ DWunion __w; \ -+ __w.ll = __umulsidi3 (u, v); \ -+ w1 = __w.s.high; \ -+ w0 = __w.s.low; \ -+} -+ -+#define count_leading_zeros(COUNT,X) ((COUNT) = __builtin_clz (X)) -+#define count_trailing_zeros(COUNT,X) ((COUNT) = __builtin_ctz (X)) -+#define COUNT_LEADING_ZEROS_0 32 -+#endif -+ - #if defined (__hppa) && W_TYPE_SIZE == 32 - #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ - __asm__ ("add %4,%5,%1\n\taddc %2,%3,%0" \ -diff -Nrup gcc-4.2.1/gcc/optabs.h gcc-4.2.1.atmel.1.3.2/gcc/optabs.h ---- gcc-4.2.1/gcc/optabs.h 2006-01-19 11:24:00.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/optabs.h 2007-09-28 10:33:07.000000000 +0200 -@@ -432,7 +432,7 @@ extern enum insn_code reload_out_optab[N - extern GTY(()) optab code_to_optab[NUM_RTX_CODE + 1]; - - --typedef rtx (*rtxfun) (rtx); -+typedef rtx (*rtxfun) (rtx, ...); - - /* Indexed by the rtx-code for a conditional (e.g. EQ, LT,...) - gives the gen_function to make a branch to test that condition. */ -diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c ---- gcc-4.2.1/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c 2005-02-19 20:48:02.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c 2007-05-07 14:24:46.000000000 +0200 -@@ -1 +1,3 @@ --/* Test no newline at eof warning when Mac line ending is used*/ /* { dg-do compile } */ int main() { return 0; } -\ No newline at end of file -+/* Test no newline at eof warning when Mac line ending is used*/ -+/* { dg-do compile } */ -+int main() { return 0; } -diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-3.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-3.c ---- gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-3.c 2005-07-20 08:39:38.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-3.c 2007-09-28 10:31:43.000000000 +0200 -@@ -5,7 +5,7 @@ - Copyright (C) 2002 Free Software Foundation Inc. - Contributed by Hans-Peter Nilsson */ - --/* { dg-do run { xfail arc-*-* avr-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ -+/* { dg-do run { xfail arc-*-* avr-*-* avr32-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ - /* { dg-options "-O2 -foptimize-sibling-calls" } */ - - /* The option -foptimize-sibling-calls is the default, but serves as -diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-4.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-4.c ---- gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-4.c 2005-07-20 08:39:38.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-4.c 2007-09-28 10:31:43.000000000 +0200 -@@ -5,7 +5,7 @@ - Copyright (C) 2002 Free Software Foundation Inc. - Contributed by Hans-Peter Nilsson */ - --/* { dg-do run { xfail arc-*-* avr-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ -+/* { dg-do run { xfail arc-*-* avr-*-* avr32-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ - /* { dg-options "-O2 -foptimize-sibling-calls" } */ - - /* The option -foptimize-sibling-calls is the default, but serves as -diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/trampoline-1.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/trampoline-1.c ---- gcc-4.2.1/gcc/testsuite/gcc.dg/trampoline-1.c 2004-08-03 10:22:26.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/trampoline-1.c 2007-09-28 10:31:43.000000000 +0200 -@@ -46,6 +46,8 @@ void foo (void) - - int main (void) - { -+#ifndef NO_TRAMPOLINES - foo (); -+#endif - return 0; - } -diff -Nrup gcc-4.2.1/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 ---- gcc-4.2.1/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 2005-05-29 18:03:43.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 2007-05-07 16:38:14.000000000 +0200 -@@ -1,4 +1,4 @@ --! { dg-do run } -+! { dg-do run } - program char_pointer_assign - ! Test character pointer assignments, required - ! to fix PR18890 and PR21297 -@@ -8,7 +8,7 @@ program char_pointer_assign - character*4, target :: t2(4) =(/"lmno","lmno","lmno","lmno"/) - character*4 :: const - character*4, pointer :: c1, c3 -- character*4, pointer :: c2(:), c4(:) -+ character*4, pointer :: c2(:), c4(:) - allocate (c3, c4(4)) - ! Scalars first. - c3 = "lmno" ! pointer = constant -@@ -24,13 +24,13 @@ program char_pointer_assign - - ! Now arrays. - c4 = "lmno" ! pointer = constant -- t2 = c4 ! target = pointer -- c2 => t2 ! pointer =>target -- const = c2(1) -+ t2 = c4 ! target = pointer -+ c2 => t2 ! pointer =>target -+ const = c2(1) - const(2:3) ="nm" ! c2(:)(2:3) = "nm" is still broken - c2 = const - c4 = c2 ! pointer = pointer -- const = c4(1) -+ const = c4(1) - const(1:1) ="o" ! c4(:)(1:1) = "o" is still broken - const(4:4) ="l" ! c4(:)(4:4) = "l" is still broken - c4 = const -diff -Nrup gcc-4.2.1/gcc/testsuite/g++.old-deja/g++.pt/static11.C gcc-4.2.1.atmel.1.3.2/gcc/testsuite/g++.old-deja/g++.pt/static11.C ---- gcc-4.2.1/gcc/testsuite/g++.old-deja/g++.pt/static11.C 2006-02-22 10:05:07.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/g++.old-deja/g++.pt/static11.C 2007-09-28 10:31:46.000000000 +0200 -@@ -2,7 +2,7 @@ - // in their dejagnu baseboard description) require that the status is - // final when exit is entered (or main returns), and not "overruled" by a - // destructor calling _exit. It's not really worth it to handle that. --// { dg-do run { xfail mmix-knuth-mmixware xtensa-*-elf* arm*-*-elf arm*-*-eabi m68k-*-elf } } -+// { dg-do run { xfail mmix-knuth-mmixware xtensa-*-elf* avr32-*-elf arm*-*-elf arm*-*-eabi m68k-*-elf } } - - // Bug: g++ was failing to destroy C::a because it was using two - // different sentry variables for construction and destruction. -diff -Nrup gcc-4.2.1/LAST_UPDATED gcc-4.2.1.atmel.1.3.2/LAST_UPDATED ---- gcc-4.2.1/LAST_UPDATED 2007-07-19 16:33:57.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/LAST_UPDATED 1970-01-01 01:00:00.000000000 +0100 -@@ -1 +0,0 @@ --Obtained from SVN: tags/gcc_4_2_1_release revision 126787 -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/copying.txt gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/copying.txt ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/copying.txt 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/copying.txt 2007-09-28 10:34:16.000000000 +0200 -@@ -1,30 +1,30 @@ --Copyright (c) 2001, Thai Open Source Software Center Ltd, Sun Microsystems. --All rights reserved. -- --Redistribution and use in source and binary forms, with or without --modification, are permitted provided that the following conditions are --met: -- -- Redistributions of source code must retain the above copyright -- notice, this list of conditions and the following disclaimer. -- -- Redistributions in binary form must reproduce the above copyright -- notice, this list of conditions and the following disclaimer in -- the documentation and/or other materials provided with the -- distribution. -- -- Neither the names of the copyright holders nor the names of its -- contributors may be used to endorse or promote products derived -- from this software without specific prior written permission. -- --THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS --"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT --LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR --A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR --CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, --EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, --PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR --PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF --LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING --NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS --SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -+Copyright (c) 2001, Thai Open Source Software Center Ltd, Sun Microsystems. -+All rights reserved. -+ -+Redistribution and use in source and binary forms, with or without -+modification, are permitted provided that the following conditions are -+met: -+ -+ Redistributions of source code must retain the above copyright -+ notice, this list of conditions and the following disclaimer. -+ -+ Redistributions in binary form must reproduce the above copyright -+ notice, this list of conditions and the following disclaimer in -+ the documentation and/or other materials provided with the -+ distribution. -+ -+ Neither the names of the copyright holders nor the names of its -+ contributors may be used to endorse or promote products derived -+ from this software without specific prior written permission. -+ -+THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -+"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -+LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -+A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR -+CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, -+EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, -+PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR -+PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF -+LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING -+NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS -+SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeBuilder.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeBuilder.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeBuilder.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeBuilder.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,45 +1,45 @@ --package org.relaxng.datatype; -- --/** -- * Creates a user-defined type by adding parameters to -- * the pre-defined type. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface DatatypeBuilder { -- -- /** -- * Adds a new parameter. -- * -- * @param name -- * The name of the parameter to be added. -- * @param strValue -- * The raw value of the parameter. Caller may not normalize -- * this value because any white space is potentially significant. -- * @param context -- * The context information which can be used by the callee to -- * acquire additional information. This context object is -- * valid only during this method call. The callee may not -- * keep a reference to this object. -- * @exception DatatypeException -- * When the given parameter is inappropriate for some reason. -- * The callee is responsible to recover from this error. -- * That is, the object should behave as if no such error -- * was occured. -- */ -- void addParameter( String name, String strValue, ValidationContext context ) -- throws DatatypeException; -- -- /** -- * Derives a new Datatype from a Datatype by parameters that -- * were already set through the addParameter method. -- * -- * @exception DatatypeException -- * DatatypeException must be thrown if the derivation is -- * somehow invalid. For example, a required parameter is missing, -- * etc. The exception should contain a diagnosis message -- * if possible. -- */ -- Datatype createDatatype() throws DatatypeException; --} -+package org.relaxng.datatype; -+ -+/** -+ * Creates a user-defined type by adding parameters to -+ * the pre-defined type. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface DatatypeBuilder { -+ -+ /** -+ * Adds a new parameter. -+ * -+ * @param name -+ * The name of the parameter to be added. -+ * @param strValue -+ * The raw value of the parameter. Caller may not normalize -+ * this value because any white space is potentially significant. -+ * @param context -+ * The context information which can be used by the callee to -+ * acquire additional information. This context object is -+ * valid only during this method call. The callee may not -+ * keep a reference to this object. -+ * @exception DatatypeException -+ * When the given parameter is inappropriate for some reason. -+ * The callee is responsible to recover from this error. -+ * That is, the object should behave as if no such error -+ * was occured. -+ */ -+ void addParameter( String name, String strValue, ValidationContext context ) -+ throws DatatypeException; -+ -+ /** -+ * Derives a new Datatype from a Datatype by parameters that -+ * were already set through the addParameter method. -+ * -+ * @exception DatatypeException -+ * DatatypeException must be thrown if the derivation is -+ * somehow invalid. For example, a required parameter is missing, -+ * etc. The exception should contain a diagnosis message -+ * if possible. -+ */ -+ Datatype createDatatype() throws DatatypeException; -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeException.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeException.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeException.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeException.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,39 +1,39 @@ --package org.relaxng.datatype; -- --/** -- * Signals Datatype related exceptions. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public class DatatypeException extends Exception { -- -- public DatatypeException( int index, String msg ) { -- super(msg); -- this.index = index; -- } -- public DatatypeException( String msg ) { -- this(UNKNOWN,msg); -- } -- /** -- * A constructor for those datatype libraries which don't support any -- * diagnostic information at all. -- */ -- public DatatypeException() { -- this(UNKNOWN,null); -- } -- -- -- private final int index; -- -- public static final int UNKNOWN = -1; -- -- /** -- * Gets the index of the content where the error occured. -- * UNKNOWN can be returned to indicate that no index information -- * is available. -- */ -- public int getIndex() { -- return index; -- } --} -+package org.relaxng.datatype; -+ -+/** -+ * Signals Datatype related exceptions. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public class DatatypeException extends Exception { -+ -+ public DatatypeException( int index, String msg ) { -+ super(msg); -+ this.index = index; -+ } -+ public DatatypeException( String msg ) { -+ this(UNKNOWN,msg); -+ } -+ /** -+ * A constructor for those datatype libraries which don't support any -+ * diagnostic information at all. -+ */ -+ public DatatypeException() { -+ this(UNKNOWN,null); -+ } -+ -+ -+ private final int index; -+ -+ public static final int UNKNOWN = -1; -+ -+ /** -+ * Gets the index of the content where the error occured. -+ * UNKNOWN can be returned to indicate that no index information -+ * is available. -+ */ -+ public int getIndex() { -+ return index; -+ } -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/Datatype.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/Datatype.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/Datatype.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/Datatype.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,237 +1,237 @@ --package org.relaxng.datatype; -- --/** -- * Datatype object. -- * -- * This object has the following functionality: -- * -- *
    -- *
  1. functionality to identify a class of character sequences. This is -- * done through the isValid method. -- * -- *
  2. functionality to produce a "value object" from a character sequence and -- * context information. -- * -- *
  3. functionality to test the equality of two value objects. -- *
-- * -- * This interface also defines the createStreamingValidator method, -- * which is intended to efficiently support the validation of -- * large character sequences. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface Datatype { -- -- /** -- * Checks if the specified 'literal' matches this Datatype -- * with respect to the current context. -- * -- * @param literal -- * the lexical representation to be checked. -- * @param context -- * If this datatype is context-dependent -- * (i.e. the {@link #isContextDependent} method returns true), -- * then the caller must provide a non-null valid context object. -- * Otherwise, the caller can pass null. -- * -- * @return -- * true if the 'literal' is a member of this Datatype; -- * false if it's not a member of this Datatype. -- */ -- boolean isValid( String literal, ValidationContext context ); -- -- /** -- * Similar to the isValid method but throws an exception with diagnosis -- * in case of errors. -- * -- *

-- * If the specified 'literal' is a valid lexical representation for this -- * datatype, then this method must return without throwing any exception. -- * If not, the callee must throw an exception (with diagnosis message, -- * if possible.) -- * -- *

-- * The application can use this method to provide detailed error message -- * to users. This method is kept separate from the isValid method to -- * achieve higher performance during normal validation. -- * -- * @exception DatatypeException -- * If the given literal is invalid, then this exception is thrown. -- * If the callee supports error diagnosis, then the exception should -- * contain a diagnosis message. -- */ -- void checkValid( String literal, ValidationContext context ) -- throws DatatypeException; -- -- /** -- * Creates an instance of a streaming validator for this type. -- * -- *

-- * By using streaming validators instead of the isValid method, -- * the caller can avoid keeping the entire string, which is -- * sometimes quite big, in memory. -- * -- * @param context -- * If this datatype is context-dependent -- * (i.e. the {@link #isContextDependent} method returns true), -- * then the caller must provide a non-null valid context object. -- * Otherwise, the caller can pass null. -- * The callee may keep a reference to this context object -- * only while the returned streaming validator is being used. -- */ -- DatatypeStreamingValidator createStreamingValidator( ValidationContext context ); -- -- /** -- * Converts lexcial value and the current context to the corresponding -- * value object. -- * -- *

-- * The caller cannot generally assume that the value object is -- * a meaningful Java object. For example, the caller cannot expect -- * this method to return java.lang.Number type for -- * the "integer" type of XML Schema Part 2. -- * -- *

-- * Also, the caller cannot assume that the equals method and -- * the hashCode method of the value object are consistent with -- * the semantics of the datatype. For that purpose, the sameValue -- * method and the valueHashCode method have to be used. Note that -- * this means you cannot use classes like -- * java.util.Hashtable to store the value objects. -- * -- *

-- * The returned value object should be used solely for the sameValue -- * and valueHashCode methods. -- * -- * @param context -- * If this datatype is context-dependent -- * (when the {@link #isContextDependent} method returns true), -- * then the caller must provide a non-null valid context object. -- * Otherwise, the caller can pass null. -- * -- * @return null -- * when the given lexical value is not a valid lexical -- * value for this type. -- */ -- Object createValue( String literal, ValidationContext context ); -- -- /** -- * Tests the equality of two value objects which were originally -- * created by the createValue method of this object. -- * -- * The behavior is undefined if objects not created by this type -- * are passed. It is the caller's responsibility to ensure that -- * value objects belong to this type. -- * -- * @return -- * true if two value objects are considered equal according to -- * the definition of this datatype; false if otherwise. -- */ -- boolean sameValue( Object value1, Object value2 ); -- -- -- /** -- * Computes the hash code for a value object, -- * which is consistent with the sameValue method. -- * -- * @return -- * hash code for the specified value object. -- */ -- int valueHashCode( Object value ); -- -- -- -- -- /** -- * Indicates that the datatype doesn't have ID/IDREF semantics. -- * -- * This value is one of the possible return values of the -- * {@link #getIdType} method. -- */ -- public static final int ID_TYPE_NULL = 0; -- -- /** -- * Indicates that RELAX NG compatibility processors should -- * treat this datatype as having ID semantics. -- * -- * This value is one of the possible return values of the -- * {@link #getIdType} method. -- */ -- public static final int ID_TYPE_ID = 1; -- -- /** -- * Indicates that RELAX NG compatibility processors should -- * treat this datatype as having IDREF semantics. -- * -- * This value is one of the possible return values of the -- * {@link #getIdType} method. -- */ -- public static final int ID_TYPE_IDREF = 2; -- -- /** -- * Indicates that RELAX NG compatibility processors should -- * treat this datatype as having IDREFS semantics. -- * -- * This value is one of the possible return values of the -- * {@link #getIdType} method. -- */ -- public static final int ID_TYPE_IDREFS = 3; -- -- /** -- * Checks if the ID/IDREF semantics is associated with this -- * datatype. -- * -- *

-- * This method is introduced to support the RELAX NG DTD -- * compatibility spec. (Of course it's always free to use -- * this method for other purposes.) -- * -- *

-- * If you are implementing a datatype library and have no idea about -- * the "RELAX NG DTD compatibility" thing, just return -- * ID_TYPE_NULL is fine. -- * -- * @return -- * If this datatype doesn't have any ID/IDREF semantics, -- * it returns {@link #ID_TYPE_NULL}. If it has such a semantics -- * (for example, XSD:ID, XSD:IDREF and comp:ID type), then -- * it returns {@link #ID_TYPE_ID}, {@link #ID_TYPE_IDREF} or -- * {@link #ID_TYPE_IDREFS}. -- */ -- public int getIdType(); -- -- -- /** -- * Checks if this datatype may need a context object for -- * the validation. -- * -- *

-- * The callee must return true even when the context -- * is not always necessary. (For example, the "QName" type -- * doesn't need a context object when validating unprefixed -- * string. But nonetheless QName must return true.) -- * -- *

-- * XSD's string and short types -- * are examples of context-independent datatypes. -- * Its QName and ENTITY types -- * are examples of context-dependent datatypes. -- * -- *

-- * When a datatype is context-independent, then -- * the {@link #isValid} method, the {@link #checkValid} method, -- * the {@link #createStreamingValidator} method and -- * the {@link #createValue} method can be called without -- * providing a context object. -- * -- * @return -- * true if this datatype is context-dependent -- * (it needs a context object sometimes); -- * -- * false if this datatype is context-independent -- * (it never needs a context object). -- */ -- public boolean isContextDependent(); --} -+package org.relaxng.datatype; -+ -+/** -+ * Datatype object. -+ * -+ * This object has the following functionality: -+ * -+ *

    -+ *
  1. functionality to identify a class of character sequences. This is -+ * done through the isValid method. -+ * -+ *
  2. functionality to produce a "value object" from a character sequence and -+ * context information. -+ * -+ *
  3. functionality to test the equality of two value objects. -+ *
-+ * -+ * This interface also defines the createStreamingValidator method, -+ * which is intended to efficiently support the validation of -+ * large character sequences. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface Datatype { -+ -+ /** -+ * Checks if the specified 'literal' matches this Datatype -+ * with respect to the current context. -+ * -+ * @param literal -+ * the lexical representation to be checked. -+ * @param context -+ * If this datatype is context-dependent -+ * (i.e. the {@link #isContextDependent} method returns true), -+ * then the caller must provide a non-null valid context object. -+ * Otherwise, the caller can pass null. -+ * -+ * @return -+ * true if the 'literal' is a member of this Datatype; -+ * false if it's not a member of this Datatype. -+ */ -+ boolean isValid( String literal, ValidationContext context ); -+ -+ /** -+ * Similar to the isValid method but throws an exception with diagnosis -+ * in case of errors. -+ * -+ *

-+ * If the specified 'literal' is a valid lexical representation for this -+ * datatype, then this method must return without throwing any exception. -+ * If not, the callee must throw an exception (with diagnosis message, -+ * if possible.) -+ * -+ *

-+ * The application can use this method to provide detailed error message -+ * to users. This method is kept separate from the isValid method to -+ * achieve higher performance during normal validation. -+ * -+ * @exception DatatypeException -+ * If the given literal is invalid, then this exception is thrown. -+ * If the callee supports error diagnosis, then the exception should -+ * contain a diagnosis message. -+ */ -+ void checkValid( String literal, ValidationContext context ) -+ throws DatatypeException; -+ -+ /** -+ * Creates an instance of a streaming validator for this type. -+ * -+ *

-+ * By using streaming validators instead of the isValid method, -+ * the caller can avoid keeping the entire string, which is -+ * sometimes quite big, in memory. -+ * -+ * @param context -+ * If this datatype is context-dependent -+ * (i.e. the {@link #isContextDependent} method returns true), -+ * then the caller must provide a non-null valid context object. -+ * Otherwise, the caller can pass null. -+ * The callee may keep a reference to this context object -+ * only while the returned streaming validator is being used. -+ */ -+ DatatypeStreamingValidator createStreamingValidator( ValidationContext context ); -+ -+ /** -+ * Converts lexcial value and the current context to the corresponding -+ * value object. -+ * -+ *

-+ * The caller cannot generally assume that the value object is -+ * a meaningful Java object. For example, the caller cannot expect -+ * this method to return java.lang.Number type for -+ * the "integer" type of XML Schema Part 2. -+ * -+ *

-+ * Also, the caller cannot assume that the equals method and -+ * the hashCode method of the value object are consistent with -+ * the semantics of the datatype. For that purpose, the sameValue -+ * method and the valueHashCode method have to be used. Note that -+ * this means you cannot use classes like -+ * java.util.Hashtable to store the value objects. -+ * -+ *

-+ * The returned value object should be used solely for the sameValue -+ * and valueHashCode methods. -+ * -+ * @param context -+ * If this datatype is context-dependent -+ * (when the {@link #isContextDependent} method returns true), -+ * then the caller must provide a non-null valid context object. -+ * Otherwise, the caller can pass null. -+ * -+ * @return null -+ * when the given lexical value is not a valid lexical -+ * value for this type. -+ */ -+ Object createValue( String literal, ValidationContext context ); -+ -+ /** -+ * Tests the equality of two value objects which were originally -+ * created by the createValue method of this object. -+ * -+ * The behavior is undefined if objects not created by this type -+ * are passed. It is the caller's responsibility to ensure that -+ * value objects belong to this type. -+ * -+ * @return -+ * true if two value objects are considered equal according to -+ * the definition of this datatype; false if otherwise. -+ */ -+ boolean sameValue( Object value1, Object value2 ); -+ -+ -+ /** -+ * Computes the hash code for a value object, -+ * which is consistent with the sameValue method. -+ * -+ * @return -+ * hash code for the specified value object. -+ */ -+ int valueHashCode( Object value ); -+ -+ -+ -+ -+ /** -+ * Indicates that the datatype doesn't have ID/IDREF semantics. -+ * -+ * This value is one of the possible return values of the -+ * {@link #getIdType} method. -+ */ -+ public static final int ID_TYPE_NULL = 0; -+ -+ /** -+ * Indicates that RELAX NG compatibility processors should -+ * treat this datatype as having ID semantics. -+ * -+ * This value is one of the possible return values of the -+ * {@link #getIdType} method. -+ */ -+ public static final int ID_TYPE_ID = 1; -+ -+ /** -+ * Indicates that RELAX NG compatibility processors should -+ * treat this datatype as having IDREF semantics. -+ * -+ * This value is one of the possible return values of the -+ * {@link #getIdType} method. -+ */ -+ public static final int ID_TYPE_IDREF = 2; -+ -+ /** -+ * Indicates that RELAX NG compatibility processors should -+ * treat this datatype as having IDREFS semantics. -+ * -+ * This value is one of the possible return values of the -+ * {@link #getIdType} method. -+ */ -+ public static final int ID_TYPE_IDREFS = 3; -+ -+ /** -+ * Checks if the ID/IDREF semantics is associated with this -+ * datatype. -+ * -+ *

-+ * This method is introduced to support the RELAX NG DTD -+ * compatibility spec. (Of course it's always free to use -+ * this method for other purposes.) -+ * -+ *

-+ * If you are implementing a datatype library and have no idea about -+ * the "RELAX NG DTD compatibility" thing, just return -+ * ID_TYPE_NULL is fine. -+ * -+ * @return -+ * If this datatype doesn't have any ID/IDREF semantics, -+ * it returns {@link #ID_TYPE_NULL}. If it has such a semantics -+ * (for example, XSD:ID, XSD:IDREF and comp:ID type), then -+ * it returns {@link #ID_TYPE_ID}, {@link #ID_TYPE_IDREF} or -+ * {@link #ID_TYPE_IDREFS}. -+ */ -+ public int getIdType(); -+ -+ -+ /** -+ * Checks if this datatype may need a context object for -+ * the validation. -+ * -+ *

-+ * The callee must return true even when the context -+ * is not always necessary. (For example, the "QName" type -+ * doesn't need a context object when validating unprefixed -+ * string. But nonetheless QName must return true.) -+ * -+ *

-+ * XSD's string and short types -+ * are examples of context-independent datatypes. -+ * Its QName and ENTITY types -+ * are examples of context-dependent datatypes. -+ * -+ *

-+ * When a datatype is context-independent, then -+ * the {@link #isValid} method, the {@link #checkValid} method, -+ * the {@link #createStreamingValidator} method and -+ * the {@link #createValue} method can be called without -+ * providing a context object. -+ * -+ * @return -+ * true if this datatype is context-dependent -+ * (it needs a context object sometimes); -+ * -+ * false if this datatype is context-independent -+ * (it never needs a context object). -+ */ -+ public boolean isContextDependent(); -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibraryFactory.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibraryFactory.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibraryFactory.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibraryFactory.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,26 +1,26 @@ --package org.relaxng.datatype; -- --/** -- * Factory class for the DatatypeLibrary class. -- * -- *

-- * The datatype library should provide the implementation of -- * this interface if it wants to be found by the schema processors. -- * The implementor also have to place a file in your jar file. -- * See the reference datatype library implementation for detail. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface DatatypeLibraryFactory --{ -- /** -- * Creates a new instance of a DatatypeLibrary that supports -- * the specified namespace URI. -- * -- * @return -- * null if the specified namespace URI is not -- * supported. -- */ -- DatatypeLibrary createDatatypeLibrary( String namespaceURI ); --} -+package org.relaxng.datatype; -+ -+/** -+ * Factory class for the DatatypeLibrary class. -+ * -+ *

-+ * The datatype library should provide the implementation of -+ * this interface if it wants to be found by the schema processors. -+ * The implementor also have to place a file in your jar file. -+ * See the reference datatype library implementation for detail. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface DatatypeLibraryFactory -+{ -+ /** -+ * Creates a new instance of a DatatypeLibrary that supports -+ * the specified namespace URI. -+ * -+ * @return -+ * null if the specified namespace URI is not -+ * supported. -+ */ -+ DatatypeLibrary createDatatypeLibrary( String namespaceURI ); -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibrary.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibrary.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibrary.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeLibrary.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,37 +1,37 @@ --package org.relaxng.datatype; -- --/** -- * A Datatype library -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface DatatypeLibrary { -- -- /** -- * Creates a new instance of DatatypeBuilder. -- * -- * The callee should throw a DatatypeException in case of an error. -- * -- * @param baseTypeLocalName -- * The local name of the base type. -- * -- * @return -- * A non-null valid datatype object. -- */ -- DatatypeBuilder createDatatypeBuilder( String baseTypeLocalName ) -- throws DatatypeException; -- -- /** -- * Gets or creates a pre-defined type. -- * -- * This is just a short-cut of -- * createDatatypeBuilder(typeLocalName).createDatatype(); -- * -- * The callee should throw a DatatypeException in case of an error. -- * -- * @return -- * A non-null valid datatype object. -- */ -- Datatype createDatatype( String typeLocalName ) throws DatatypeException; --} -+package org.relaxng.datatype; -+ -+/** -+ * A Datatype library -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface DatatypeLibrary { -+ -+ /** -+ * Creates a new instance of DatatypeBuilder. -+ * -+ * The callee should throw a DatatypeException in case of an error. -+ * -+ * @param baseTypeLocalName -+ * The local name of the base type. -+ * -+ * @return -+ * A non-null valid datatype object. -+ */ -+ DatatypeBuilder createDatatypeBuilder( String baseTypeLocalName ) -+ throws DatatypeException; -+ -+ /** -+ * Gets or creates a pre-defined type. -+ * -+ * This is just a short-cut of -+ * createDatatypeBuilder(typeLocalName).createDatatype(); -+ * -+ * The callee should throw a DatatypeException in case of an error. -+ * -+ * @return -+ * A non-null valid datatype object. -+ */ -+ Datatype createDatatype( String typeLocalName ) throws DatatypeException; -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeStreamingValidator.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeStreamingValidator.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeStreamingValidator.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/DatatypeStreamingValidator.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,46 +1,46 @@ --package org.relaxng.datatype; -- --/** -- * Datatype streaming validator. -- * -- *

-- * The streaming validator is an optional feature that is useful for -- * certain Datatypes. It allows the caller to incrementally provide -- * the literal. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface DatatypeStreamingValidator { -- -- /** -- * Passes an additional fragment of the literal. -- * -- *

-- * The application can call this method several times, then call -- * the isValid method (or the checkValid method) to check the validity -- * of the accumulated characters. -- */ -- void addCharacters( char[] buf, int start, int len ); -- -- /** -- * Tells if the accumulated literal is valid with respect to -- * the underlying Datatype. -- * -- * @return -- * True if it is valid. False if otherwise. -- */ -- boolean isValid(); -- -- /** -- * Similar to the isValid method, but this method throws -- * Exception (with possibly diagnostic information), instead of -- * returning false. -- * -- * @exception DatatypeException -- * If the callee supports the diagnosis and the accumulated -- * literal is invalid, then this exception that possibly -- * contains diagnosis information is thrown. -- */ -- void checkValid() throws DatatypeException; --} -+package org.relaxng.datatype; -+ -+/** -+ * Datatype streaming validator. -+ * -+ *

-+ * The streaming validator is an optional feature that is useful for -+ * certain Datatypes. It allows the caller to incrementally provide -+ * the literal. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface DatatypeStreamingValidator { -+ -+ /** -+ * Passes an additional fragment of the literal. -+ * -+ *

-+ * The application can call this method several times, then call -+ * the isValid method (or the checkValid method) to check the validity -+ * of the accumulated characters. -+ */ -+ void addCharacters( char[] buf, int start, int len ); -+ -+ /** -+ * Tells if the accumulated literal is valid with respect to -+ * the underlying Datatype. -+ * -+ * @return -+ * True if it is valid. False if otherwise. -+ */ -+ boolean isValid(); -+ -+ /** -+ * Similar to the isValid method, but this method throws -+ * Exception (with possibly diagnostic information), instead of -+ * returning false. -+ * -+ * @exception DatatypeException -+ * If the callee supports the diagnosis and the accumulated -+ * literal is invalid, then this exception that possibly -+ * contains diagnosis information is thrown. -+ */ -+ void checkValid() throws DatatypeException; -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/DatatypeLibraryLoader.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/DatatypeLibraryLoader.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/DatatypeLibraryLoader.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/DatatypeLibraryLoader.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,262 +1,262 @@ --/** -- * Copyright (c) 2001, Thai Open Source Software Center Ltd -- * All rights reserved. -- * -- * Redistribution and use in source and binary forms, with or without -- * modification, are permitted provided that the following conditions are -- * met: -- * -- * Redistributions of source code must retain the above copyright -- * notice, this list of conditions and the following disclaimer. -- * -- * Redistributions in binary form must reproduce the above copyright -- * notice, this list of conditions and the following disclaimer in -- * the documentation and/or other materials provided with the -- * distribution. -- * -- * Neither the name of the Thai Open Source Software Center Ltd nor -- * the names of its contributors may be used to endorse or promote -- * products derived from this software without specific prior written -- * permission. -- * -- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -- * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -- * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -- * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR -- * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, -- * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, -- * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR -- * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF -- * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING -- * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS -- * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -- */ --package org.relaxng.datatype.helpers; -- --import org.relaxng.datatype.DatatypeLibraryFactory; --import org.relaxng.datatype.DatatypeLibrary; --import java.util.Enumeration; --import java.util.NoSuchElementException; --import java.util.Vector; --import java.io.Reader; --import java.io.InputStream; --import java.io.InputStreamReader; --import java.io.BufferedReader; --import java.io.IOException; --import java.io.UnsupportedEncodingException; --import java.net.URL; -- --/** -- * Discovers the datatype library implementation from the classpath. -- * -- *

-- * The call of the createDatatypeLibrary method finds an implementation -- * from a given datatype library URI at run-time. -- */ --public class DatatypeLibraryLoader implements DatatypeLibraryFactory { -- private final Service service = new Service(DatatypeLibraryFactory.class); -- -- public DatatypeLibrary createDatatypeLibrary(String uri) { -- for (Enumeration e = service.getProviders(); -- e.hasMoreElements();) { -- DatatypeLibraryFactory factory -- = (DatatypeLibraryFactory)e.nextElement(); -- DatatypeLibrary library = factory.createDatatypeLibrary(uri); -- if (library != null) -- return library; -- } -- return null; -- } -- -- private static class Service { -- private final Class serviceClass; -- private final Enumeration configFiles; -- private Enumeration classNames = null; -- private final Vector providers = new Vector(); -- private Loader loader; -- -- private class ProviderEnumeration implements Enumeration { -- private int nextIndex = 0; -- -- public boolean hasMoreElements() { -- return nextIndex < providers.size() || moreProviders(); -- } -- -- public Object nextElement() { -- try { -- return providers.elementAt(nextIndex++); -- } -- catch (ArrayIndexOutOfBoundsException e) { -- throw new NoSuchElementException(); -- } -- } -- } -- -- private static class Singleton implements Enumeration { -- private Object obj; -- private Singleton(Object obj) { -- this.obj = obj; -- } -- -- public boolean hasMoreElements() { -- return obj != null; -- } -- -- public Object nextElement() { -- if (obj == null) -- throw new NoSuchElementException(); -- Object tem = obj; -- obj = null; -- return tem; -- } -- } -- -- // JDK 1.1 -- private static class Loader { -- Enumeration getResources(String resName) { -- ClassLoader cl = Loader.class.getClassLoader(); -- URL url; -- if (cl == null) -- url = ClassLoader.getSystemResource(resName); -- else -- url = cl.getResource(resName); -- return new Singleton(url); -- } -- -- Class loadClass(String name) throws ClassNotFoundException { -- return Class.forName(name); -- } -- } -- -- // JDK 1.2+ -- private static class Loader2 extends Loader { -- private ClassLoader cl; -- -- Loader2() { -- cl = Loader2.class.getClassLoader(); -- // If the thread context class loader has the class loader -- // of this class as an ancestor, use the thread context class -- // loader. Otherwise, the thread context class loader -- // probably hasn't been set up properly, so don't use it. -- ClassLoader clt = Thread.currentThread().getContextClassLoader(); -- for (ClassLoader tem = clt; tem != null; tem = tem.getParent()) -- if (tem == cl) { -- cl = clt; -- break; -- } -- } -- -- Enumeration getResources(String resName) { -- try { -- return cl.getResources(resName); -- } -- catch (IOException e) { -- return new Singleton(null); -- } -- } -- -- Class loadClass(String name) throws ClassNotFoundException { -- return Class.forName(name, true, cl); -- } -- } -- -- public Service(Class cls) { -- try { -- loader = new Loader2(); -- } -- catch (NoSuchMethodError e) { -- loader = new Loader(); -- } -- serviceClass = cls; -- String resName = "META-INF/services/" + serviceClass.getName(); -- configFiles = loader.getResources(resName); -- } -- -- public Enumeration getProviders() { -- return new ProviderEnumeration(); -- } -- -- synchronized private boolean moreProviders() { -- for (;;) { -- while (classNames == null) { -- if (!configFiles.hasMoreElements()) -- return false; -- classNames = parseConfigFile((URL)configFiles.nextElement()); -- } -- while (classNames.hasMoreElements()) { -- String className = (String)classNames.nextElement(); -- try { -- Class cls = loader.loadClass(className); -- Object obj = cls.newInstance(); -- if (serviceClass.isInstance(obj)) { -- providers.addElement(obj); -- return true; -- } -- } -- catch (ClassNotFoundException e) { } -- catch (InstantiationException e) { } -- catch (IllegalAccessException e) { } -- catch (LinkageError e) { } -- } -- classNames = null; -- } -- } -- -- private static final int START = 0; -- private static final int IN_NAME = 1; -- private static final int IN_COMMENT = 2; -- -- private static Enumeration parseConfigFile(URL url) { -- try { -- InputStream in = url.openStream(); -- Reader r; -- try { -- r = new InputStreamReader(in, "UTF-8"); -- } -- catch (UnsupportedEncodingException e) { -- r = new InputStreamReader(in, "UTF8"); -- } -- r = new BufferedReader(r); -- Vector tokens = new Vector(); -- StringBuffer tokenBuf = new StringBuffer(); -- int state = START; -- for (;;) { -- int n = r.read(); -- if (n < 0) -- break; -- char c = (char)n; -- switch (c) { -- case '\r': -- case '\n': -- state = START; -- break; -- case ' ': -- case '\t': -- break; -- case '#': -- state = IN_COMMENT; -- break; -- default: -- if (state != IN_COMMENT) { -- state = IN_NAME; -- tokenBuf.append(c); -- } -- break; -- } -- if (tokenBuf.length() != 0 && state != IN_NAME) { -- tokens.addElement(tokenBuf.toString()); -- tokenBuf.setLength(0); -- } -- } -- if (tokenBuf.length() != 0) -- tokens.addElement(tokenBuf.toString()); -- return tokens.elements(); -- } -- catch (IOException e) { -- return null; -- } -- } -- } -- --} -- -+/** -+ * Copyright (c) 2001, Thai Open Source Software Center Ltd -+ * All rights reserved. -+ * -+ * Redistribution and use in source and binary forms, with or without -+ * modification, are permitted provided that the following conditions are -+ * met: -+ * -+ * Redistributions of source code must retain the above copyright -+ * notice, this list of conditions and the following disclaimer. -+ * -+ * Redistributions in binary form must reproduce the above copyright -+ * notice, this list of conditions and the following disclaimer in -+ * the documentation and/or other materials provided with the -+ * distribution. -+ * -+ * Neither the name of the Thai Open Source Software Center Ltd nor -+ * the names of its contributors may be used to endorse or promote -+ * products derived from this software without specific prior written -+ * permission. -+ * -+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR -+ * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, -+ * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, -+ * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR -+ * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF -+ * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING -+ * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS -+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -+ */ -+package org.relaxng.datatype.helpers; -+ -+import org.relaxng.datatype.DatatypeLibraryFactory; -+import org.relaxng.datatype.DatatypeLibrary; -+import java.util.Enumeration; -+import java.util.NoSuchElementException; -+import java.util.Vector; -+import java.io.Reader; -+import java.io.InputStream; -+import java.io.InputStreamReader; -+import java.io.BufferedReader; -+import java.io.IOException; -+import java.io.UnsupportedEncodingException; -+import java.net.URL; -+ -+/** -+ * Discovers the datatype library implementation from the classpath. -+ * -+ *

-+ * The call of the createDatatypeLibrary method finds an implementation -+ * from a given datatype library URI at run-time. -+ */ -+public class DatatypeLibraryLoader implements DatatypeLibraryFactory { -+ private final Service service = new Service(DatatypeLibraryFactory.class); -+ -+ public DatatypeLibrary createDatatypeLibrary(String uri) { -+ for (Enumeration e = service.getProviders(); -+ e.hasMoreElements();) { -+ DatatypeLibraryFactory factory -+ = (DatatypeLibraryFactory)e.nextElement(); -+ DatatypeLibrary library = factory.createDatatypeLibrary(uri); -+ if (library != null) -+ return library; -+ } -+ return null; -+ } -+ -+ private static class Service { -+ private final Class serviceClass; -+ private final Enumeration configFiles; -+ private Enumeration classNames = null; -+ private final Vector providers = new Vector(); -+ private Loader loader; -+ -+ private class ProviderEnumeration implements Enumeration { -+ private int nextIndex = 0; -+ -+ public boolean hasMoreElements() { -+ return nextIndex < providers.size() || moreProviders(); -+ } -+ -+ public Object nextElement() { -+ try { -+ return providers.elementAt(nextIndex++); -+ } -+ catch (ArrayIndexOutOfBoundsException e) { -+ throw new NoSuchElementException(); -+ } -+ } -+ } -+ -+ private static class Singleton implements Enumeration { -+ private Object obj; -+ private Singleton(Object obj) { -+ this.obj = obj; -+ } -+ -+ public boolean hasMoreElements() { -+ return obj != null; -+ } -+ -+ public Object nextElement() { -+ if (obj == null) -+ throw new NoSuchElementException(); -+ Object tem = obj; -+ obj = null; -+ return tem; -+ } -+ } -+ -+ // JDK 1.1 -+ private static class Loader { -+ Enumeration getResources(String resName) { -+ ClassLoader cl = Loader.class.getClassLoader(); -+ URL url; -+ if (cl == null) -+ url = ClassLoader.getSystemResource(resName); -+ else -+ url = cl.getResource(resName); -+ return new Singleton(url); -+ } -+ -+ Class loadClass(String name) throws ClassNotFoundException { -+ return Class.forName(name); -+ } -+ } -+ -+ // JDK 1.2+ -+ private static class Loader2 extends Loader { -+ private ClassLoader cl; -+ -+ Loader2() { -+ cl = Loader2.class.getClassLoader(); -+ // If the thread context class loader has the class loader -+ // of this class as an ancestor, use the thread context class -+ // loader. Otherwise, the thread context class loader -+ // probably hasn't been set up properly, so don't use it. -+ ClassLoader clt = Thread.currentThread().getContextClassLoader(); -+ for (ClassLoader tem = clt; tem != null; tem = tem.getParent()) -+ if (tem == cl) { -+ cl = clt; -+ break; -+ } -+ } -+ -+ Enumeration getResources(String resName) { -+ try { -+ return cl.getResources(resName); -+ } -+ catch (IOException e) { -+ return new Singleton(null); -+ } -+ } -+ -+ Class loadClass(String name) throws ClassNotFoundException { -+ return Class.forName(name, true, cl); -+ } -+ } -+ -+ public Service(Class cls) { -+ try { -+ loader = new Loader2(); -+ } -+ catch (NoSuchMethodError e) { -+ loader = new Loader(); -+ } -+ serviceClass = cls; -+ String resName = "META-INF/services/" + serviceClass.getName(); -+ configFiles = loader.getResources(resName); -+ } -+ -+ public Enumeration getProviders() { -+ return new ProviderEnumeration(); -+ } -+ -+ synchronized private boolean moreProviders() { -+ for (;;) { -+ while (classNames == null) { -+ if (!configFiles.hasMoreElements()) -+ return false; -+ classNames = parseConfigFile((URL)configFiles.nextElement()); -+ } -+ while (classNames.hasMoreElements()) { -+ String className = (String)classNames.nextElement(); -+ try { -+ Class cls = loader.loadClass(className); -+ Object obj = cls.newInstance(); -+ if (serviceClass.isInstance(obj)) { -+ providers.addElement(obj); -+ return true; -+ } -+ } -+ catch (ClassNotFoundException e) { } -+ catch (InstantiationException e) { } -+ catch (IllegalAccessException e) { } -+ catch (LinkageError e) { } -+ } -+ classNames = null; -+ } -+ } -+ -+ private static final int START = 0; -+ private static final int IN_NAME = 1; -+ private static final int IN_COMMENT = 2; -+ -+ private static Enumeration parseConfigFile(URL url) { -+ try { -+ InputStream in = url.openStream(); -+ Reader r; -+ try { -+ r = new InputStreamReader(in, "UTF-8"); -+ } -+ catch (UnsupportedEncodingException e) { -+ r = new InputStreamReader(in, "UTF8"); -+ } -+ r = new BufferedReader(r); -+ Vector tokens = new Vector(); -+ StringBuffer tokenBuf = new StringBuffer(); -+ int state = START; -+ for (;;) { -+ int n = r.read(); -+ if (n < 0) -+ break; -+ char c = (char)n; -+ switch (c) { -+ case '\r': -+ case '\n': -+ state = START; -+ break; -+ case ' ': -+ case '\t': -+ break; -+ case '#': -+ state = IN_COMMENT; -+ break; -+ default: -+ if (state != IN_COMMENT) { -+ state = IN_NAME; -+ tokenBuf.append(c); -+ } -+ break; -+ } -+ if (tokenBuf.length() != 0 && state != IN_NAME) { -+ tokens.addElement(tokenBuf.toString()); -+ tokenBuf.setLength(0); -+ } -+ } -+ if (tokenBuf.length() != 0) -+ tokens.addElement(tokenBuf.toString()); -+ return tokens.elements(); -+ } -+ catch (IOException e) { -+ return null; -+ } -+ } -+ } -+ -+} -+ -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/ParameterlessDatatypeBuilder.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/ParameterlessDatatypeBuilder.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/ParameterlessDatatypeBuilder.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/ParameterlessDatatypeBuilder.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,42 +1,42 @@ --package org.relaxng.datatype.helpers; -- --import org.relaxng.datatype.*; -- --/** -- * Dummy implementation of {@link DatatypeBuilder}. -- * -- * This implementation can be used for Datatypes which have no parameters. -- * Any attempt to add parameters will be rejected. -- * -- *

-- * Typical usage would be: -- *


-- * class MyDatatypeLibrary implements DatatypeLibrary {
-- *     ....
-- *     DatatypeBuilder createDatatypeBuilder( String typeName ) {
-- *         return new ParameterleessDatatypeBuilder(createDatatype(typeName));
-- *     }
-- *     ....
-- * }
-- * 
-- * -- * @author Kohsuke KAWAGUCHI -- */ --public final class ParameterlessDatatypeBuilder implements DatatypeBuilder { -- -- /** This type object is returned for the derive method. */ -- private final Datatype baseType; -- -- public ParameterlessDatatypeBuilder( Datatype baseType ) { -- this.baseType = baseType; -- } -- -- public void addParameter( String name, String strValue, ValidationContext context ) -- throws DatatypeException { -- throw new DatatypeException(); -- } -- -- public Datatype createDatatype() throws DatatypeException { -- return baseType; -- } --} -+package org.relaxng.datatype.helpers; -+ -+import org.relaxng.datatype.*; -+ -+/** -+ * Dummy implementation of {@link DatatypeBuilder}. -+ * -+ * This implementation can be used for Datatypes which have no parameters. -+ * Any attempt to add parameters will be rejected. -+ * -+ *

-+ * Typical usage would be: -+ *


-+ * class MyDatatypeLibrary implements DatatypeLibrary {
-+ *     ....
-+ *     DatatypeBuilder createDatatypeBuilder( String typeName ) {
-+ *         return new ParameterleessDatatypeBuilder(createDatatype(typeName));
-+ *     }
-+ *     ....
-+ * }
-+ * 
-+ * -+ * @author Kohsuke KAWAGUCHI -+ */ -+public final class ParameterlessDatatypeBuilder implements DatatypeBuilder { -+ -+ /** This type object is returned for the derive method. */ -+ private final Datatype baseType; -+ -+ public ParameterlessDatatypeBuilder( Datatype baseType ) { -+ this.baseType = baseType; -+ } -+ -+ public void addParameter( String name, String strValue, ValidationContext context ) -+ throws DatatypeException { -+ throw new DatatypeException(); -+ } -+ -+ public Datatype createDatatype() throws DatatypeException { -+ return baseType; -+ } -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/StreamingValidatorImpl.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/StreamingValidatorImpl.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/StreamingValidatorImpl.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/helpers/StreamingValidatorImpl.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,55 +1,55 @@ --package org.relaxng.datatype.helpers; -- --import org.relaxng.datatype.*; -- --/** -- * Dummy implementation of {@link DatatypeStreamingValidator}. -- * -- *

-- * This implementation can be used as a quick hack when the performance -- * of streaming validation is not important. And this implementation -- * also shows you how to implement the DatatypeStreamingValidator interface. -- * -- *

-- * Typical usage would be: -- *


-- * class MyDatatype implements Datatype {
-- *     ....
-- *     public DatatypeStreamingValidator createStreamingValidator( ValidationContext context ) {
-- *         return new StreamingValidatorImpl(this,context);
-- *     }
-- *     ....
-- * }
-- * 
-- * -- * @author Kohsuke KAWAGUCHI -- */ --public final class StreamingValidatorImpl implements DatatypeStreamingValidator { -- -- /** This buffer accumulates characters. */ -- private final StringBuffer buffer = new StringBuffer(); -- -- /** Datatype obejct that creates this streaming validator. */ -- private final Datatype baseType; -- -- /** The current context. */ -- private final ValidationContext context; -- -- public void addCharacters( char[] buf, int start, int len ) { -- // append characters to the current buffer. -- buffer.append(buf,start,len); -- } -- -- public boolean isValid() { -- return baseType.isValid(buffer.toString(),context); -- } -- -- public void checkValid() throws DatatypeException { -- baseType.checkValid(buffer.toString(),context); -- } -- -- public StreamingValidatorImpl( Datatype baseType, ValidationContext context ) { -- this.baseType = baseType; -- this.context = context; -- } --} -+package org.relaxng.datatype.helpers; -+ -+import org.relaxng.datatype.*; -+ -+/** -+ * Dummy implementation of {@link DatatypeStreamingValidator}. -+ * -+ *

-+ * This implementation can be used as a quick hack when the performance -+ * of streaming validation is not important. And this implementation -+ * also shows you how to implement the DatatypeStreamingValidator interface. -+ * -+ *

-+ * Typical usage would be: -+ *


-+ * class MyDatatype implements Datatype {
-+ *     ....
-+ *     public DatatypeStreamingValidator createStreamingValidator( ValidationContext context ) {
-+ *         return new StreamingValidatorImpl(this,context);
-+ *     }
-+ *     ....
-+ * }
-+ * 
-+ * -+ * @author Kohsuke KAWAGUCHI -+ */ -+public final class StreamingValidatorImpl implements DatatypeStreamingValidator { -+ -+ /** This buffer accumulates characters. */ -+ private final StringBuffer buffer = new StringBuffer(); -+ -+ /** Datatype obejct that creates this streaming validator. */ -+ private final Datatype baseType; -+ -+ /** The current context. */ -+ private final ValidationContext context; -+ -+ public void addCharacters( char[] buf, int start, int len ) { -+ // append characters to the current buffer. -+ buffer.append(buf,start,len); -+ } -+ -+ public boolean isValid() { -+ return baseType.isValid(buffer.toString(),context); -+ } -+ -+ public void checkValid() throws DatatypeException { -+ baseType.checkValid(buffer.toString(),context); -+ } -+ -+ public StreamingValidatorImpl( Datatype baseType, ValidationContext context ) { -+ this.baseType = baseType; -+ this.context = context; -+ } -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/ValidationContext.java gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/ValidationContext.java ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/ValidationContext.java 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/org/relaxng/datatype/ValidationContext.java 2007-09-28 10:34:16.000000000 +0200 -@@ -1,66 +1,66 @@ --package org.relaxng.datatype; -- --/** -- * An interface that must be implemented by caller to -- * provide context information that is necessary to -- * perform validation of some Datatypes. -- * -- * @author James Clark -- * @author Kohsuke KAWAGUCHI -- */ --public interface ValidationContext { -- -- /** -- * Resolves a namespace prefix to the corresponding namespace URI. -- * -- * This method is used for validating the QName type, for example. -- * -- *

-- * If the prefix is "" (empty string), it indicates -- * an unprefixed value. The callee -- * should resolve it as for an unprefixed -- * element, rather than for an unprefixed attribute. -- * -- *

-- * If the prefix is "xml", then the callee must resolve -- * this prefix into "http://www.w3.org/XML/1998/namespace", -- * as defined in the XML Namespaces Recommendation. -- * -- * @return -- * namespace URI of this prefix. -- * If the specified prefix is not declared, -- * the implementation must return null. -- */ -- String resolveNamespacePrefix( String prefix ); -- -- /** -- * Returns the base URI of the context. The null string may be returned -- * if no base URI is known. -- */ -- String getBaseUri(); -- -- /** -- * Checks if an unparsed entity is declared with the -- * specified name. -- * -- * @return -- * true -- * if the DTD has an unparsed entity declaration for -- * the specified name. -- * false -- * otherwise. -- */ -- boolean isUnparsedEntity( String entityName ); -- -- /** -- * Checks if a notation is declared with the -- * specified name. -- * -- * @return -- * true -- * if the DTD has a notation declaration for the specified name. -- * false -- * otherwise. -- */ -- boolean isNotation( String notationName ); --} -+package org.relaxng.datatype; -+ -+/** -+ * An interface that must be implemented by caller to -+ * provide context information that is necessary to -+ * perform validation of some Datatypes. -+ * -+ * @author James Clark -+ * @author Kohsuke KAWAGUCHI -+ */ -+public interface ValidationContext { -+ -+ /** -+ * Resolves a namespace prefix to the corresponding namespace URI. -+ * -+ * This method is used for validating the QName type, for example. -+ * -+ *

-+ * If the prefix is "" (empty string), it indicates -+ * an unprefixed value. The callee -+ * should resolve it as for an unprefixed -+ * element, rather than for an unprefixed attribute. -+ * -+ *

-+ * If the prefix is "xml", then the callee must resolve -+ * this prefix into "http://www.w3.org/XML/1998/namespace", -+ * as defined in the XML Namespaces Recommendation. -+ * -+ * @return -+ * namespace URI of this prefix. -+ * If the specified prefix is not declared, -+ * the implementation must return null. -+ */ -+ String resolveNamespacePrefix( String prefix ); -+ -+ /** -+ * Returns the base URI of the context. The null string may be returned -+ * if no base URI is known. -+ */ -+ String getBaseUri(); -+ -+ /** -+ * Checks if an unparsed entity is declared with the -+ * specified name. -+ * -+ * @return -+ * true -+ * if the DTD has an unparsed entity declaration for -+ * the specified name. -+ * false -+ * otherwise. -+ */ -+ boolean isUnparsedEntity( String entityName ); -+ -+ /** -+ * Checks if a notation is declared with the -+ * specified name. -+ * -+ * @return -+ * true -+ * if the DTD has a notation declaration for the specified name. -+ * false -+ * otherwise. -+ */ -+ boolean isNotation( String notationName ); -+} -diff -Nrup gcc-4.2.1/libjava/classpath/external/relaxngDatatype/README.txt gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/README.txt ---- gcc-4.2.1/libjava/classpath/external/relaxngDatatype/README.txt 2006-03-10 14:25:35.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libjava/classpath/external/relaxngDatatype/README.txt 2007-09-28 10:34:16.000000000 +0200 -@@ -1,54 +1,54 @@ --====================================================================== -- README FILE FOR DATATYPE INTERFACES FOR RELAX NG --====================================================================== -- -- -- --RELAX NG supports multiple datatype vocabularies. To achive this, an --interface between datatype vocabularies and schema processors is --necessary. This interface is intended to be a standard Java interface --for this purpose. -- -- ------------------------------------------------------------------------ --LICENSE ------------------------------------------------------------------------ -- --See copying.txt. -- --Note: this license is the BSD license. -- -- -- ------------------------------------------------------------------------ --FOR DEVELOPER ------------------------------------------------------------------------ -- --If you are planning to implement a datatype library, A sample datatype --library implementation by James Clark is available at [1], which --comes with documentation and source code. -- --If you are planning to implement a schema processor, then don't forget --to check out org.relaxng.datatype.helpers.DatatypeLibraryLoader, as --this allows you to dynamically locate datatype implementations. -- -- ------------------------------------------------------------------------ --LINKS ------------------------------------------------------------------------ -- --* OASIS RELAX NG TC -- http://www.oasis-open.org/committees/relax-ng/ --* RELAX home page -- http://www.xml.gr.jp/relax/ -- -- ------------------------------------------------------------------------ --REFERENCES ------------------------------------------------------------------------ --[1] Sample datatype library implementation by James Clark -- http://www.thaiopensource.com/relaxng/datatype-sample.zip -- --Document written by Kohsuke Kawaguchi (kohsuke.kawaguchi@sun.com) --====================================================================== --END OF README -+====================================================================== -+ README FILE FOR DATATYPE INTERFACES FOR RELAX NG -+====================================================================== -+ -+ -+ -+RELAX NG supports multiple datatype vocabularies. To achive this, an -+interface between datatype vocabularies and schema processors is -+necessary. This interface is intended to be a standard Java interface -+for this purpose. -+ -+ -+---------------------------------------------------------------------- -+LICENSE -+---------------------------------------------------------------------- -+ -+See copying.txt. -+ -+Note: this license is the BSD license. -+ -+ -+ -+---------------------------------------------------------------------- -+FOR DEVELOPER -+---------------------------------------------------------------------- -+ -+If you are planning to implement a datatype library, A sample datatype -+library implementation by James Clark is available at [1], which -+comes with documentation and source code. -+ -+If you are planning to implement a schema processor, then don't forget -+to check out org.relaxng.datatype.helpers.DatatypeLibraryLoader, as -+this allows you to dynamically locate datatype implementations. -+ -+ -+---------------------------------------------------------------------- -+LINKS -+---------------------------------------------------------------------- -+ -+* OASIS RELAX NG TC -+ http://www.oasis-open.org/committees/relax-ng/ -+* RELAX home page -+ http://www.xml.gr.jp/relax/ -+ -+ -+---------------------------------------------------------------------- -+REFERENCES -+---------------------------------------------------------------------- -+[1] Sample datatype library implementation by James Clark -+ http://www.thaiopensource.com/relaxng/datatype-sample.zip -+ -+Document written by Kohsuke Kawaguchi (kohsuke.kawaguchi@sun.com) -+====================================================================== -+END OF README -diff -Nrup gcc-4.2.1/libstdc++-v3/acinclude.m4 gcc-4.2.1.atmel.1.3.2/libstdc++-v3/acinclude.m4 ---- gcc-4.2.1/libstdc++-v3/acinclude.m4 2007-06-29 01:02:05.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/acinclude.m4 2007-09-28 10:33:34.000000000 +0200 -@@ -125,15 +125,6 @@ AC_DEFUN([GLIBCXX_CONFIGURE], [ - ## other macros from doing the same. This should be automated.) -pme - need_libmath=no - -- # Check for uClibc since Linux platforms use different configuration -- # directories depending on the C library in use. -- AC_EGREP_CPP([_using_uclibc], [ -- #include -- #if __UCLIBC__ -- _using_uclibc -- #endif -- ], uclibc=yes, uclibc=no) -- - # Find platform-specific directories containing configuration info. - # Also possibly modify flags used elsewhere, as needed by the platform. - GLIBCXX_CHECK_HOST -@@ -1389,8 +1380,8 @@ AC_DEFUN([GLIBCXX_ENABLE_CLOCALE], [ - #endif - int main() - { -- const char __one[] = "Äuglein Augmen"; -- const char __two[] = "Äuglein"; -+ const char __one[] = "Äuglein Augmen"; -+ const char __two[] = "Äuglein"; - int i; - int j; - __locale_t loc; -diff -Nrup gcc-4.2.1/libstdc++-v3/config/os/gnu-linux/ctype_base.h gcc-4.2.1.atmel.1.3.2/libstdc++-v3/config/os/gnu-linux/ctype_base.h ---- gcc-4.2.1/libstdc++-v3/config/os/gnu-linux/ctype_base.h 2006-12-01 13:56:23.000000000 +0100 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/config/os/gnu-linux/ctype_base.h 2007-09-28 10:33:32.000000000 +0200 -@@ -31,6 +31,8 @@ - // - // ISO C++ 14882: 22.1 Locales - // -+#include -+#include - - /** @file ctype_base.h - * This is an internal header file, included by other library headers. -@@ -45,8 +47,12 @@ _GLIBCXX_BEGIN_NAMESPACE(std) - struct ctype_base - { - // Non-standard typedefs. -- typedef const int* __to_type; -- -+#ifdef __UCLIBC__ -+ typedef const __ctype_touplow_t* __to_type; -+#else -+ typedef const int* __to_type; -+#endif -+ - // NB: Offsets into ctype::_M_table force a particular size - // on the mask type. Because of this, we don't use an enum. - typedef unsigned short mask; -diff -Nrup gcc-4.2.1/libstdc++-v3/include/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/include/Makefile.in ---- gcc-4.2.1/libstdc++-v3/include/Makefile.in 2007-07-05 13:46:00.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/include/Makefile.in 2007-09-28 10:33:21.000000000 +0200 -@@ -36,6 +36,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ - $(top_srcdir)/fragment.am - subdir = include -diff -Nrup gcc-4.2.1/libstdc++-v3/libmath/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libmath/Makefile.in ---- gcc-4.2.1/libstdc++-v3/libmath/Makefile.in 2006-10-16 21:08:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libmath/Makefile.in 2007-09-28 10:33:16.000000000 +0200 -@@ -37,6 +37,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - subdir = libmath - DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in - ACLOCAL_M4 = $(top_srcdir)/aclocal.m4 -diff -Nrup gcc-4.2.1/libstdc++-v3/libsupc++/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libsupc++/Makefile.in ---- gcc-4.2.1/libstdc++-v3/libsupc++/Makefile.in 2006-10-16 21:08:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libsupc++/Makefile.in 2007-09-28 10:33:21.000000000 +0200 -@@ -38,6 +38,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - DIST_COMMON = $(glibcxxinstall_HEADERS) $(srcdir)/Makefile.am \ - $(srcdir)/Makefile.in $(top_srcdir)/fragment.am - subdir = libsupc++ -diff -Nrup gcc-4.2.1/libstdc++-v3/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/Makefile.in ---- gcc-4.2.1/libstdc++-v3/Makefile.in 2006-10-16 21:08:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/Makefile.in 2007-09-28 10:33:34.000000000 +0200 -@@ -36,6 +36,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - DIST_COMMON = README $(am__configure_deps) $(srcdir)/../config.guess \ - $(srcdir)/../config.sub $(srcdir)/../install-sh \ - $(srcdir)/../ltmain.sh $(srcdir)/../missing \ -diff -Nrup gcc-4.2.1/libstdc++-v3/po/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/po/Makefile.in ---- gcc-4.2.1/libstdc++-v3/po/Makefile.in 2006-10-16 21:08:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/po/Makefile.in 2007-09-28 10:33:34.000000000 +0200 -@@ -36,6 +36,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ - $(top_srcdir)/fragment.am - subdir = po -diff -Nrup gcc-4.2.1/libstdc++-v3/src/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/src/Makefile.in ---- gcc-4.2.1/libstdc++-v3/src/Makefile.in 2006-10-16 21:08:22.000000000 +0200 -+++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/src/Makefile.in 2007-09-28 10:33:21.000000000 +0200 -@@ -36,6 +36,7 @@ POST_UNINSTALL = : - build_triplet = @build@ - host_triplet = @host@ - target_triplet = @target@ -+LIBOBJDIR = - DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ - $(top_srcdir)/fragment.am - subdir = src diff --git a/toolchain/gcc/4.2.1/901-avr32-add-mno-pic.patch b/toolchain/gcc/4.2.1/901-avr32-add-mno-pic.patch deleted file mode 100644 index c21450a9ad..0000000000 --- a/toolchain/gcc/4.2.1/901-avr32-add-mno-pic.patch +++ /dev/null @@ -1,25 +0,0 @@ -Index: gcc/gcc/config/avr32/avr32.opt -=================================================================== ---- gcc/gcc/config/avr32/avr32.opt (revision 31583) -+++ gcc/gcc/config/avr32/avr32.opt (working copy) -@@ -71,3 +71,6 @@ - Target Report Var(avr32_imm_in_const_pool) Init(-1) - Put large immediates in constant pool. This is enabled by default for archs with insn-cache. - -+mno-pic -+Target Report RejectNegative Mask(NO_PIC) -+Do not generate position-independent code. -Index: gcc/gcc/config/avr32/avr32.c -=================================================================== ---- gcc/gcc/config/avr32/avr32.c (revision 31583) -+++ gcc/gcc/config/avr32/avr32.c (working copy) -@@ -230,6 +230,9 @@ - avr32_imm_in_const_pool = 0; - } - -+ if (TARGET_NO_PIC) -+ flag_pic = 0; -+ - avr32_add_gc_roots (); - } - diff --git a/toolchain/gcc/4.2.1/902-avr32-fix-pic-redefinition-warning.patch b/toolchain/gcc/4.2.1/902-avr32-fix-pic-redefinition-warning.patch deleted file mode 100644 index ae9fa5d188..0000000000 --- a/toolchain/gcc/4.2.1/902-avr32-fix-pic-redefinition-warning.patch +++ /dev/null @@ -1,16 +0,0 @@ -Index: gcc/gcc/config/avr32/linux-elf.h -=================================================================== ---- gcc/gcc/config/avr32/linux-elf.h (revision 31583) -+++ gcc/gcc/config/avr32/linux-elf.h (working copy) -@@ -122,11 +122,6 @@ - builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ - if (TARGET_FAST_FLOAT) \ - builtin_define ("__AVR32_FAST_FLOAT__"); \ -- if (flag_pic) \ -- { \ -- builtin_define ("__PIC__"); \ -- builtin_define ("__pic__"); \ -- } \ - } \ - while (0) - diff --git a/toolchain/gcc/4.2.1/903-avr32-fix-expanding-builtin-xchg.patch b/toolchain/gcc/4.2.1/903-avr32-fix-expanding-builtin-xchg.patch deleted file mode 100644 index 4a4ac60932..0000000000 --- a/toolchain/gcc/4.2.1/903-avr32-fix-expanding-builtin-xchg.patch +++ /dev/null @@ -1,12 +0,0 @@ ---- gcc-4.2.1.orig/gcc/config/avr32/avr32.c 2007-10-05 07:52:51.000000000 +0200 -+++ gcc-4.2.1/gcc/config/avr32/avr32.c 2007-10-05 07:54:49.000000000 +0200 -@@ -1388,7 +1388,8 @@ avr32_expand_builtin (tree exp, - op1 = copy_to_mode_reg (mode1, op1); - } - -- op0 = gen_rtx_MEM (SImode, op0); -+ op0 = force_reg (GET_MODE (op0), op0); -+ op0 = gen_rtx_MEM (GET_MODE (op0), op0); - if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) - { - error diff --git a/toolchain/gcc/4.2.1/905-avr32-fix-folding-machine-reorg-optimizations.patch b/toolchain/gcc/4.2.1/905-avr32-fix-folding-machine-reorg-optimizations.patch deleted file mode 100644 index 7a160d6e95..0000000000 --- a/toolchain/gcc/4.2.1/905-avr32-fix-folding-machine-reorg-optimizations.patch +++ /dev/null @@ -1,48 +0,0 @@ -Index: a/gcc/config/avr32/avr32.c -=================================================================== ---- a/gcc/config/avr32/avr32.c (revision 31997) -+++ b/gcc/config/avr32/avr32.c (working copy) -@@ -6323,25 +6323,26 @@ - /* If used any other place than as a pointer or as the - destination register we failed */ - if (!(single_set (scan) -- && GET_CODE (PATTERN (scan)) == SET -- && ((MEM_P (SET_DEST (PATTERN (scan))) -- && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) -- && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == -- REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) -- && -- REG_P (XEXP -- (SET_SRC (PATTERN (scan)), -- 0)) -- && -- REGNO (XEXP -- (SET_SRC (PATTERN (scan)), -- 0)) == REGNO (reg)))) -- && !(GET_CODE (PATTERN (scan)) == SET -- && REG_P (SET_DEST (PATTERN (scan))) -- && !regno_use_in (REGNO (reg), -- SET_SRC (PATTERN (scan))))) -+ && GET_CODE (PATTERN (scan)) == SET -+ && ((MEM_P (SET_DEST (PATTERN (scan))) -+ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) -+ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == REGNO (reg)) -+ || (MEM_P (SET_SRC (PATTERN (scan))) -+ && REG_P (XEXP (SET_SRC (PATTERN (scan)), 0)) -+ && REGNO (XEXP -+ (SET_SRC (PATTERN (scan)), 0)) == REGNO (reg)))) -+ && !(GET_CODE (PATTERN (scan)) == SET -+ && REG_P (SET_DEST (PATTERN (scan))) -+ && !regno_use_in (REGNO (reg), -+ SET_SRC (PATTERN (scan))))) - break; - -+ /* We cannot replace the pointer in TImode insns -+ as these has a differene addressing mode than the other -+ memory insns. */ -+ if ( GET_MODE (SET_DEST (PATTERN (scan))) == TImode ) -+ break; -+ - /* Check if register is dead or set in this insn */ - if (dead_or_set_p (scan, reg)) - { diff --git a/toolchain/gcc/4.2.1/906-avr32-use-rjmp-instead-of-got-when-jumping.patch b/toolchain/gcc/4.2.1/906-avr32-use-rjmp-instead-of-got-when-jumping.patch deleted file mode 100644 index 5d3c8d18a3..0000000000 --- a/toolchain/gcc/4.2.1/906-avr32-use-rjmp-instead-of-got-when-jumping.patch +++ /dev/null @@ -1,85 +0,0 @@ -Index: a/gcc/config/avr32/avr32.c -=================================================================== ---- a/gcc/config/avr32/avr32.c (revision 32101) -+++ b/gcc/config/avr32/avr32.c (working copy) -@@ -695,8 +695,7 @@ - - - if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") -- || vcall_offset -- || flag_pic) -+ || vcall_offset) - { - fputs ("\tpushm\tlr\n", file); - } -@@ -728,47 +727,23 @@ - } - - -- if ( (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") -- || vcall_offset) -- && !flag_pic ) -+ if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") -+ || vcall_offset) - { - fputs ("\tpopm\tlr\n", file); - } -- -- if (flag_pic) -- { -- /* Load the got into lr and then load the pointer -- to the function from the got and put it on the stack. -- We can then call the function and restore lr by issuing -- a doubleword load from the stack. We do not use a popm/ldm -- since it will be treated as a return and might need a flushing -- of the return-stack if available. */ -- rtx label = gen_label_rtx (); -- /* Load the got. */ -- fputs ("\tlddpc\tlr, 0f\n", file); -- (*targetm.asm_out.internal_label) (file, "L", -- CODE_LABEL_NUMBER (label)); -- fputs ("\trsub\tlr, pc\n", file); -- /* Load the function pointer. */ -- fputs ("\tld.w\tlr, lr[", file); -- assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); -- fputs ("@got]\n", file); -- /* Push the function pointer on the stack.*/ -- fputs ("\tpushm\tlr\n", file); -- /* Restore the old lr value and load the function pointer into -- pc. */ -- fputs ("\tld.d\tlr,sp++\n", file); -- fprintf (file, "\t.align 2\n"); -- fprintf (file, "0:\t.long\t.L%d - _GLOBAL_OFFSET_TABLE_\n", CODE_LABEL_NUMBER (label)); -- } -- else -- { -- fprintf (file, "\tlddpc\tpc, 0f\n"); -- fprintf (file, "\t.align 2\n"); -- fputs ("0:\t.long\t", file); -- assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); -- fputc ('\n', file); -- } -+ -+ /* Jump to the function. We assume that we can use an rjmp since the -+ function to jump to is local and probably not too far away from -+ the thunk. If this assumption proves to be wrong we could implement -+ this jump by calculating the offset between the jump source and destination -+ and put this in the constant pool and then perform an add to pc. -+ This would also be legitimate PIC code. But for now we hope that an rjmp -+ will be sufficient... -+ */ -+ fputs ("\trjmp\t", file); -+ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); -+ fputc ('\n', file); - } - - /* Implements target hook vector_mode_supported. */ -@@ -1742,7 +1717,7 @@ - if (TREE_CODE (*node) != FUNCTION_DECL) - { - warning ("`%s' attribute only applies to functions", -- IDENTIFIER_POINTER (name)); -+ IDENTIFIER_POINTER (name)); - *no_add_attrs = true; - } - /* FIXME: the argument if any is checked for type attributes; should it diff --git a/toolchain/gcc/4.2.1/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch b/toolchain/gcc/4.2.1/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch deleted file mode 100644 index 63ead2be79..0000000000 --- a/toolchain/gcc/4.2.1/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch +++ /dev/null @@ -1,14 +0,0 @@ -diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.c gcc-4.2.1-new/gcc/config/avr32/avr32.c ---- gcc-4.2.1/gcc/config/avr32/avr32.c 2007-11-15 18:20:25.000000000 +0100 -+++ gcc-4.2.1-new/gcc/config/avr32/avr32.c 2007-11-16 13:34:39.000000000 +0100 -@@ -705,8 +705,8 @@ avr32_output_mi_thunk (FILE * file, - { - if (avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21")) - { -- fprintf (file, "\tsub\t%s, -0x%x\n", reg_names[this_regno], -- mi_delta); -+ fprintf (file, "\tsub\t%s, %d\n", reg_names[this_regno], -+ -mi_delta); - } - else - { diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/900-avr32-atmel.1.3.2.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/900-avr32-atmel.1.3.2.patch new file mode 100644 index 0000000000..0f3da28503 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/900-avr32-atmel.1.3.2.patch @@ -0,0 +1,20631 @@ +diff -Nrup gcc-4.2.1/configure.in gcc-4.2.1.atmel.1.3.2/configure.in +--- gcc-4.2.1/configure.in 2007-05-30 15:48:07.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/configure.in 2007-09-28 10:34:36.000000000 +0200 +@@ -503,6 +503,9 @@ case "${target}" in + arm-*-riscix*) + noconfigdirs="$noconfigdirs ld target-libgloss ${libgcj}" + ;; ++ avr32-*-*) ++ noconfigdirs="$noconfigdirs target-libiberty target-libmudflap target-libffi ${libgcj}" ++ ;; + avr-*-*) + noconfigdirs="$noconfigdirs target-libiberty target-libstdc++-v3 ${libgcj}" + ;; +diff -Nrup gcc-4.2.1/gcc/builtins.c gcc-4.2.1.atmel.1.3.2/gcc/builtins.c +--- gcc-4.2.1/gcc/builtins.c 2007-03-29 18:19:32.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/builtins.c 2007-09-28 10:33:08.000000000 +0200 +@@ -9223,7 +9223,7 @@ validate_arglist (tree arglist, ...) + + do + { +- code = va_arg (ap, enum tree_code); ++ code = va_arg (ap, int); + switch (code) + { + case 0: +diff -Nrup gcc-4.2.1/gcc/calls.c gcc-4.2.1.atmel.1.3.2/gcc/calls.c +--- gcc-4.2.1/gcc/calls.c 2007-06-20 08:44:26.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/calls.c 2007-09-28 10:33:09.000000000 +0200 +@@ -3448,7 +3448,7 @@ emit_library_call_value_1 (int retval, r + for (; count < nargs; count++) + { + rtx val = va_arg (p, rtx); +- enum machine_mode mode = va_arg (p, enum machine_mode); ++ enum machine_mode mode = va_arg (p, int); + + /* We cannot convert the arg value to the mode the library wants here; + must do it earlier where we know the signedness of the arg. */ +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.c gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.c +--- gcc-4.2.1/gcc/config/avr32/avr32.c 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.c 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,7060 @@ ++/* ++ Target hooks and helper functions for AVR32. ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ Initial porting by Anders �dland. ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++#include "config.h" ++#include "system.h" ++#include "coretypes.h" ++#include "tm.h" ++#include "rtl.h" ++#include "tree.h" ++#include "obstack.h" ++#include "regs.h" ++#include "hard-reg-set.h" ++#include "real.h" ++#include "insn-config.h" ++#include "conditions.h" ++#include "output.h" ++#include "insn-attr.h" ++#include "flags.h" ++#include "reload.h" ++#include "function.h" ++#include "expr.h" ++#include "optabs.h" ++#include "toplev.h" ++#include "recog.h" ++#include "ggc.h" ++#include "except.h" ++#include "c-pragma.h" ++#include "integrate.h" ++#include "tm_p.h" ++#include "langhooks.h" ++ ++#include "target.h" ++#include "target-def.h" ++ ++#include ++ ++/* Forward definitions of types. */ ++typedef struct minipool_node Mnode; ++typedef struct minipool_fixup Mfix; ++ ++/* Obstack for minipool constant handling. */ ++static struct obstack minipool_obstack; ++static char *minipool_startobj; ++static rtx minipool_vector_label; ++ ++/* True if we are currently building a constant table. */ ++int making_const_table; ++ ++/* Some forward function declarations */ ++static unsigned long avr32_isr_value (tree); ++static unsigned long avr32_compute_func_type (void); ++static tree avr32_handle_isr_attribute (tree *, tree, tree, int, bool *); ++static tree avr32_handle_acall_attribute (tree *, tree, tree, int, bool *); ++static tree avr32_handle_fndecl_attribute (tree * node, tree name, tree args, ++ int flags, bool * no_add_attrs); ++static void avr32_reorg (void); ++bool avr32_return_in_msb (tree type); ++bool avr32_vector_mode_supported (enum machine_mode mode); ++static void avr32_init_libfuncs (void); ++ ++ ++static void ++avr32_add_gc_roots (void) ++ { ++ gcc_obstack_init (&minipool_obstack); ++ minipool_startobj = (char *) obstack_alloc (&minipool_obstack, 0); ++ } ++ ++ ++/* List of all known AVR32 parts */ ++static const struct part_type_s avr32_part_types[] = { ++ /* name, part_type, architecture type, macro */ ++ {"none", PART_TYPE_AVR32_NONE, ARCH_TYPE_AVR32_AP, "__AVR32__"}, ++ {"ap7000", PART_TYPE_AVR32_AP7000, ARCH_TYPE_AVR32_AP, "__AVR32_AP7000__"}, ++ {"ap7010", PART_TYPE_AVR32_AP7010, ARCH_TYPE_AVR32_AP, "__AVR32_AP7010__"}, ++ {"ap7020", PART_TYPE_AVR32_AP7020, ARCH_TYPE_AVR32_AP, "__AVR32_AP7020__"}, ++ {"uc3a0256", PART_TYPE_AVR32_UC3A0256, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3A0256__"}, ++ {"uc3a0512", PART_TYPE_AVR32_UC3A0512, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3A0512__"}, ++ {"uc3a1128", PART_TYPE_AVR32_UC3A1128, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3A1128__"}, ++ {"uc3a1256", PART_TYPE_AVR32_UC3A1256, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3A1256__"}, ++ {"uc3a1512", PART_TYPE_AVR32_UC3A1512, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3A1512__"}, ++ {"uc3b064", PART_TYPE_AVR32_UC3B064, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B064__"}, ++ {"uc3b0128", PART_TYPE_AVR32_UC3B0128, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B0128__"}, ++ {"uc3b0256", PART_TYPE_AVR32_UC3B0256, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B0256__"}, ++ {"uc3b164", PART_TYPE_AVR32_UC3B164, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B164__"}, ++ {"uc3b1128", PART_TYPE_AVR32_UC3B1128, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B1128__"}, ++ {"uc3b1256", PART_TYPE_AVR32_UC3B1256, ARCH_TYPE_AVR32_UC, ++ "__AVR32_UC3B1256__"}, ++ {NULL, 0, 0, NULL} ++}; ++ ++/* List of all known AVR32 architectures */ ++static const struct arch_type_s avr32_arch_types[] = { ++ /* name, architecture type, microarchitecture type, feature flags, macro */ ++ {"ap", ARCH_TYPE_AVR32_AP, UARCH_TYPE_AVR32B, ++ (FLAG_AVR32_HAS_DSP ++ | FLAG_AVR32_HAS_SIMD ++ | FLAG_AVR32_HAS_UNALIGNED_WORD ++ | FLAG_AVR32_HAS_CACHES ++ | FLAG_AVR32_HAS_BRANCH_PRED ++ | FLAG_AVR32_HAS_RETURN_STACK), ++ "__AVR32_AP__"}, ++ {"uc", ARCH_TYPE_AVR32_UC, UARCH_TYPE_AVR32A, ++ (FLAG_AVR32_HAS_DSP | FLAG_AVR32_HAS_RMW), ++ "__AVR32_UC__"}, ++ {NULL, 0, 0, 0, NULL} ++}; ++ ++/* Default arch name */ ++const char *avr32_arch_name = "ap"; ++const char *avr32_part_name = "none"; ++ ++const struct part_type_s *avr32_part; ++const struct arch_type_s *avr32_arch; ++ ++/* Set default target_flags. */ ++#undef TARGET_DEFAULT_TARGET_FLAGS ++#define TARGET_DEFAULT_TARGET_FLAGS \ ++ (MASK_HAS_ASM_ADDR_PSEUDOS | MASK_MD_REORG_OPTIMIZATION) ++ ++void ++avr32_optimization_options (int level, ++ int size){ ++ if (AVR32_ALWAYS_PIC) ++ flag_pic = 1; ++ ++ /* Enable section anchors if optimization is enabled. */ ++ if (level > 0 || size) ++ flag_section_anchors = 1; ++} ++ ++/* Override command line options */ ++void ++avr32_override_options (void) ++ { ++ const struct part_type_s *part; ++ const struct arch_type_s *arch; ++ ++ /* Check if part type is set. */ ++ for (part = avr32_part_types; part->name; part++) ++ if (strcmp (part->name, avr32_part_name) == 0) ++ break; ++ ++ avr32_part = part; ++ ++ if (!part->name) ++ { ++ fprintf (stderr, "Unknown part `%s' specified\nKnown part names:\n", ++ avr32_part_name); ++ for (part = avr32_part_types; part->name; part++) ++ fprintf (stderr, "\t%s\n", part->name); ++ avr32_part = &avr32_part_types[PART_TYPE_AVR32_NONE]; ++ } ++ ++ avr32_arch = &avr32_arch_types[avr32_part->arch_type]; ++ ++ /* If part was set to "none" then check if arch was set. */ ++ if (strcmp (avr32_part->name, "none") == 0) ++ { ++ /* Check if arch type is set. */ ++ for (arch = avr32_arch_types; arch->name; arch++) ++ if (strcmp (arch->name, avr32_arch_name) == 0) ++ break; ++ ++ avr32_arch = arch; ++ ++ if (!arch->name) ++ { ++ fprintf (stderr, "Unknown arch `%s' specified\nKnown arch names:\n", ++ avr32_arch_name); ++ for (arch = avr32_arch_types; arch->name; arch++) ++ fprintf (stderr, "\t%s\n", arch->name); ++ avr32_arch = &avr32_arch_types[ARCH_TYPE_AVR32_AP]; ++ } ++ } ++ ++ /* If optimization level is two or greater, then align start of loops to a ++ word boundary since this will allow folding the first insn of the loop. ++ Do this only for targets supporting branch prediction. */ ++ if (optimize >= 2 && TARGET_BRANCH_PRED) ++ align_loops = 2; ++ ++ ++ /* Enable section anchors if optimization is enabled. */ ++ if (optimize > 0 || optimize_size) ++ flag_section_anchors = 1; ++ ++ /* Enable fast-float library if unsafe math optimizations ++ are used. */ ++ if (flag_unsafe_math_optimizations) ++ target_flags |= MASK_FAST_FLOAT; ++ ++ /* Check if we should set avr32_imm_in_const_pool ++ based on if caches are present or not. */ ++ if ( avr32_imm_in_const_pool == -1 ) ++ { ++ if ( TARGET_CACHES ) ++ avr32_imm_in_const_pool = 1; ++ else ++ avr32_imm_in_const_pool = 0; ++ } ++ ++ avr32_add_gc_roots (); ++ } ++ ++ ++/* ++If defined, a function that outputs the assembler code for entry to a ++function. The prologue is responsible for setting up the stack frame, ++initializing the frame pointer register, saving registers that must be ++saved, and allocating size additional bytes of storage for the ++local variables. size is an integer. file is a stdio ++stream to which the assembler code should be output. ++ ++The label for the beginning of the function need not be output by this ++macro. That has already been done when the macro is run. ++ ++To determine which registers to save, the macro can refer to the array ++regs_ever_live: element r is nonzero if hard register ++r is used anywhere within the function. This implies the function ++prologue should save register r, provided it is not one of the ++call-used registers. (TARGET_ASM_FUNCTION_EPILOGUE must likewise use ++regs_ever_live.) ++ ++On machines that have ``register windows'', the function entry code does ++not save on the stack the registers that are in the windows, even if ++they are supposed to be preserved by function calls; instead it takes ++appropriate steps to ``push'' the register stack, if any non-call-used ++registers are used in the function. ++ ++On machines where functions may or may not have frame-pointers, the ++function entry code must vary accordingly; it must set up the frame ++pointer if one is wanted, and not otherwise. To determine whether a ++frame pointer is in wanted, the macro can refer to the variable ++frame_pointer_needed. The variable's value will be 1 at run ++time in a function that needs a frame pointer. (see Elimination). ++ ++The function entry code is responsible for allocating any stack space ++required for the function. This stack space consists of the regions ++listed below. In most cases, these regions are allocated in the ++order listed, with the last listed region closest to the top of the ++stack (the lowest address if STACK_GROWS_DOWNWARD is defined, and ++the highest address if it is not defined). You can use a different order ++for a machine if doing so is more convenient or required for ++compatibility reasons. Except in cases where required by standard ++or by a debugger, there is no reason why the stack layout used by GCC ++need agree with that used by other compilers for a machine. ++ */ ++ ++#undef TARGET_ASM_FUNCTION_PROLOGUE ++#define TARGET_ASM_FUNCTION_PROLOGUE avr32_target_asm_function_prologue ++ ++ ++#undef TARGET_DEFAULT_SHORT_ENUMS ++#define TARGET_DEFAULT_SHORT_ENUMS hook_bool_void_false ++ ++#undef TARGET_PROMOTE_FUNCTION_ARGS ++#define TARGET_PROMOTE_FUNCTION_ARGS hook_bool_tree_true ++ ++#undef TARGET_PROMOTE_FUNCTION_RETURN ++#define TARGET_PROMOTE_FUNCTION_RETURN hook_bool_tree_true ++ ++#undef TARGET_PROMOTE_PROTOTYPES ++#define TARGET_PROMOTE_PROTOTYPES hook_bool_tree_true ++ ++#undef TARGET_MUST_PASS_IN_STACK ++#define TARGET_MUST_PASS_IN_STACK avr32_must_pass_in_stack ++ ++#undef TARGET_PASS_BY_REFERENCE ++#define TARGET_PASS_BY_REFERENCE avr32_pass_by_reference ++ ++#undef TARGET_STRICT_ARGUMENT_NAMING ++#define TARGET_STRICT_ARGUMENT_NAMING avr32_strict_argument_naming ++ ++#undef TARGET_VECTOR_MODE_SUPPORTED_P ++#define TARGET_VECTOR_MODE_SUPPORTED_P avr32_vector_mode_supported ++ ++#undef TARGET_RETURN_IN_MEMORY ++#define TARGET_RETURN_IN_MEMORY avr32_return_in_memory ++ ++#undef TARGET_RETURN_IN_MSB ++#define TARGET_RETURN_IN_MSB avr32_return_in_msb ++ ++#undef TARGET_ARG_PARTIAL_BYTES ++#define TARGET_ARG_PARTIAL_BYTES avr32_arg_partial_bytes ++ ++#undef TARGET_STRIP_NAME_ENCODING ++#define TARGET_STRIP_NAME_ENCODING avr32_strip_name_encoding ++ ++#define streq(string1, string2) (strcmp (string1, string2) == 0) ++ ++#undef TARGET_NARROW_VOLATILE_BITFIELD ++#define TARGET_NARROW_VOLATILE_BITFIELD hook_bool_void_false ++ ++#undef TARGET_ATTRIBUTE_TABLE ++#define TARGET_ATTRIBUTE_TABLE avr32_attribute_table ++ ++#undef TARGET_COMP_TYPE_ATTRIBUTES ++#define TARGET_COMP_TYPE_ATTRIBUTES avr32_comp_type_attributes ++ ++ ++#undef TARGET_RTX_COSTS ++#define TARGET_RTX_COSTS avr32_rtx_costs ++ ++#undef TARGET_CANNOT_FORCE_CONST_MEM ++#define TARGET_CANNOT_FORCE_CONST_MEM avr32_cannot_force_const_mem ++ ++#undef TARGET_ASM_INTEGER ++#define TARGET_ASM_INTEGER avr32_assemble_integer ++ ++#undef TARGET_FUNCTION_VALUE ++#define TARGET_FUNCTION_VALUE avr32_function_value ++ ++#undef TARGET_MIN_ANCHOR_OFFSET ++#define TARGET_MIN_ANCHOR_OFFSET (0) ++ ++#undef TARGET_MAX_ANCHOR_OFFSET ++#define TARGET_MAX_ANCHOR_OFFSET ((1 << 15) - 1) ++ ++ ++/* ++ * Switches to the appropriate section for output of constant pool ++ * entry x in mode. You can assume that x is some kind of constant in ++ * RTL. The argument mode is redundant except in the case of a ++ * const_int rtx. Select the section by calling readonly_data_ section ++ * or one of the alternatives for other sections. align is the ++ * constant alignment in bits. ++ * ++ * The default version of this function takes care of putting symbolic ++ * constants in flag_ pic mode in data_section and everything else in ++ * readonly_data_section. ++ */ ++//#undef TARGET_ASM_SELECT_RTX_SECTION ++//#define TARGET_ASM_SELECT_RTX_SECTION avr32_select_rtx_section ++ ++ ++/* ++ * If non-null, this hook performs a target-specific pass over the ++ * instruction stream. The compiler will run it at all optimization ++ * levels, just before the point at which it normally does ++ * delayed-branch scheduling. ++ * ++ * The exact purpose of the hook varies from target to target. Some ++ * use it to do transformations that are necessary for correctness, ++ * such as laying out in-function constant pools or avoiding hardware ++ * hazards. Others use it as an opportunity to do some ++ * machine-dependent optimizations. ++ * ++ * You need not implement the hook if it has nothing to do. The ++ * default definition is null. ++ */ ++#undef TARGET_MACHINE_DEPENDENT_REORG ++#define TARGET_MACHINE_DEPENDENT_REORG avr32_reorg ++ ++/* Target hook for assembling integer objects. ++ Need to handle integer vectors */ ++static bool ++avr32_assemble_integer (rtx x, unsigned int size, int aligned_p) ++ { ++ if (avr32_vector_mode_supported (GET_MODE (x))) ++ { ++ int i, units; ++ ++ if (GET_CODE (x) != CONST_VECTOR) ++ abort (); ++ ++ units = CONST_VECTOR_NUNITS (x); ++ ++ switch (GET_MODE (x)) ++ { ++ case V2HImode: ++ size = 2; ++ break; ++ case V4QImode: ++ size = 1; ++ break; ++ default: ++ abort (); ++ } ++ ++ for (i = 0; i < units; i++) ++ { ++ rtx elt; ++ ++ elt = CONST_VECTOR_ELT (x, i); ++ assemble_integer (elt, size, i == 0 ? 32 : size * BITS_PER_UNIT, 1); ++ } ++ ++ return true; ++ } ++ ++ return default_assemble_integer (x, size, aligned_p); ++ } ++ ++/* ++ * This target hook describes the relative costs of RTL expressions. ++ * ++ * The cost may depend on the precise form of the expression, which is ++ * available for examination in x, and the rtx code of the expression ++ * in which it is contained, found in outer_code. code is the ++ * expression code--redundant, since it can be obtained with GET_CODE ++ * (x). ++ * ++ * In implementing this hook, you can use the construct COSTS_N_INSNS ++ * (n) to specify a cost equal to n fast instructions. ++ * ++ * On entry to the hook, *total contains a default estimate for the ++ * cost of the expression. The hook should modify this value as ++ * necessary. Traditionally, the default costs are COSTS_N_INSNS (5) ++ * for multiplications, COSTS_N_INSNS (7) for division and modulus ++ * operations, and COSTS_N_INSNS (1) for all other operations. ++ * ++ * When optimizing for code size, i.e. when optimize_size is non-zero, ++ * this target hook should be used to estimate the relative size cost ++ * of an expression, again relative to COSTS_N_INSNS. ++ * ++ * The hook returns true when all subexpressions of x have been ++ * processed, and false when rtx_cost should recurse. ++ */ ++ ++/* Worker routine for avr32_rtx_costs. */ ++static inline int ++avr32_rtx_costs_1 (rtx x, enum rtx_code code ATTRIBUTE_UNUSED, ++ enum rtx_code outer ATTRIBUTE_UNUSED) ++ { ++ enum machine_mode mode = GET_MODE (x); ++ ++ switch (GET_CODE (x)) ++ { ++ case MEM: ++ /* Using pre decrement / post increment memory operations on the ++ avr32_uc architecture means that two writebacks must be performed ++ and hence two cycles are needed. */ ++ if (!optimize_size ++ && GET_MODE_SIZE (mode) <= 2 * UNITS_PER_WORD ++ && avr32_arch->arch_type == ARCH_TYPE_AVR32_UC ++ && (GET_CODE (XEXP (x, 0)) == PRE_DEC ++ || GET_CODE (XEXP (x, 0)) == POST_INC)) ++ return COSTS_N_INSNS (5); ++ ++ /* Memory costs quite a lot for the first word, but subsequent words ++ load at the equivalent of a single insn each. */ ++ if (GET_MODE_SIZE (mode) > UNITS_PER_WORD) ++ return COSTS_N_INSNS (3 + (GET_MODE_SIZE (mode) / UNITS_PER_WORD)); ++ ++ return COSTS_N_INSNS (4); ++ case SYMBOL_REF: ++ case CONST: ++ /* These are valid for the pseudo insns: lda.w and call which operates ++ on direct addresses. We assume that the cost of a lda.w is the same ++ as the cost of a ld.w insn. */ ++ return (outer == SET) ? COSTS_N_INSNS (4) : COSTS_N_INSNS (1); ++ case DIV: ++ case MOD: ++ case UDIV: ++ case UMOD: ++ return optimize_size ? COSTS_N_INSNS (1) : COSTS_N_INSNS (16); ++ ++ case ROTATE: ++ case ROTATERT: ++ if (mode == TImode) ++ return COSTS_N_INSNS (100); ++ ++ if (mode == DImode) ++ return COSTS_N_INSNS (10); ++ return COSTS_N_INSNS (4); ++ case ASHIFT: ++ case LSHIFTRT: ++ case ASHIFTRT: ++ case NOT: ++ if (mode == TImode) ++ return COSTS_N_INSNS (10); ++ ++ if (mode == DImode) ++ return COSTS_N_INSNS (4); ++ return COSTS_N_INSNS (1); ++ case PLUS: ++ case MINUS: ++ case NEG: ++ case COMPARE: ++ case ABS: ++ if (GET_MODE_CLASS (mode) == MODE_FLOAT) ++ return COSTS_N_INSNS (100); ++ ++ if (mode == TImode) ++ return COSTS_N_INSNS (50); ++ ++ if (mode == DImode) ++ return COSTS_N_INSNS (2); ++ return COSTS_N_INSNS (1); ++ ++ case MULT: ++ { ++ if (GET_MODE_CLASS (mode) == MODE_FLOAT) ++ return COSTS_N_INSNS (300); ++ ++ if (mode == TImode) ++ return COSTS_N_INSNS (16); ++ ++ if (mode == DImode) ++ return COSTS_N_INSNS (4); ++ ++ if (mode == HImode) ++ return COSTS_N_INSNS (2); ++ ++ return COSTS_N_INSNS (3); ++ } ++ case IF_THEN_ELSE: ++ if (GET_CODE (XEXP (x, 1)) == PC || GET_CODE (XEXP (x, 2)) == PC) ++ return COSTS_N_INSNS (4); ++ return COSTS_N_INSNS (1); ++ case SIGN_EXTEND: ++ case ZERO_EXTEND: ++ /* Sign/Zero extensions of registers cost quite much since these ++ instrcutions only take one register operand which means that gcc ++ often must insert some move instrcutions */ ++ if (mode == QImode || mode == HImode) ++ return (COSTS_N_INSNS (GET_CODE (XEXP (x, 0)) == MEM ? 0 : 1)); ++ return COSTS_N_INSNS (4); ++ case UNSPEC: ++ /* divmod operations */ ++ if (XINT (x, 1) == UNSPEC_UDIVMODSI4_INTERNAL ++ || XINT (x, 1) == UNSPEC_DIVMODSI4_INTERNAL) ++ { ++ return optimize_size ? COSTS_N_INSNS (1) : COSTS_N_INSNS (16); ++ } ++ /* Fallthrough */ ++ default: ++ return COSTS_N_INSNS (1); ++ } ++ } ++ ++static bool ++avr32_rtx_costs (rtx x, int code, int outer_code, int *total) ++ { ++ *total = avr32_rtx_costs_1 (x, code, outer_code); ++ return true; ++ } ++ ++ ++bool ++avr32_cannot_force_const_mem (rtx x ATTRIBUTE_UNUSED) ++ { ++ /* Do not want symbols in the constant pool when compiling pic or if using ++ address pseudo instructions. */ ++ return ((flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS) ++ && avr32_find_symbol (x) != NULL_RTX); ++ } ++ ++ ++/* Table of machine attributes. */ ++const struct attribute_spec avr32_attribute_table[] = { ++ /* { name, min_len, max_len, decl_req, type_req, fn_type_req, handler } */ ++ /* Interrupt Service Routines have special prologue and epilogue ++ requirements. */ ++ {"isr", 0, 1, false, false, false, avr32_handle_isr_attribute}, ++ {"interrupt", 0, 1, false, false, false, avr32_handle_isr_attribute}, ++ {"acall", 0, 1, false, true, true, avr32_handle_acall_attribute}, ++ {"naked", 0, 0, true, false, false, avr32_handle_fndecl_attribute}, ++ {NULL, 0, 0, false, false, false, NULL} ++}; ++ ++ ++typedef struct ++{ ++ const char *const arg; ++ const unsigned long return_value; ++} ++isr_attribute_arg; ++ ++static const isr_attribute_arg isr_attribute_args[] = { ++ {"FULL", AVR32_FT_ISR_FULL}, ++ {"full", AVR32_FT_ISR_FULL}, ++ {"HALF", AVR32_FT_ISR_HALF}, ++ {"half", AVR32_FT_ISR_HALF}, ++ {"NONE", AVR32_FT_ISR_NONE}, ++ {"none", AVR32_FT_ISR_NONE}, ++ {"UNDEF", AVR32_FT_ISR_NONE}, ++ {"undef", AVR32_FT_ISR_NONE}, ++ {"SWI", AVR32_FT_ISR_NONE}, ++ {"swi", AVR32_FT_ISR_NONE}, ++ {NULL, AVR32_FT_ISR_NONE} ++}; ++ ++/* Returns the (interrupt) function type of the current ++ function, or AVR32_FT_UNKNOWN if the type cannot be determined. */ ++ ++static unsigned long ++avr32_isr_value (tree argument) ++ { ++ const isr_attribute_arg *ptr; ++ const char *arg; ++ ++ /* No argument - default to ISR_NONE. */ ++ if (argument == NULL_TREE) ++ return AVR32_FT_ISR_NONE; ++ ++ /* Get the value of the argument. */ ++ if (TREE_VALUE (argument) == NULL_TREE ++ || TREE_CODE (TREE_VALUE (argument)) != STRING_CST) ++ return AVR32_FT_UNKNOWN; ++ ++ arg = TREE_STRING_POINTER (TREE_VALUE (argument)); ++ ++ /* Check it against the list of known arguments. */ ++ for (ptr = isr_attribute_args; ptr->arg != NULL; ptr++) ++ if (streq (arg, ptr->arg)) ++ return ptr->return_value; ++ ++ /* An unrecognized interrupt type. */ ++ return AVR32_FT_UNKNOWN; ++ } ++ ++ ++ ++/* ++These hooks specify assembly directives for creating certain kinds ++of integer object. The TARGET_ASM_BYTE_OP directive creates a ++byte-sized object, the TARGET_ASM_ALIGNED_HI_OP one creates an ++aligned two-byte object, and so on. Any of the hooks may be ++NULL, indicating that no suitable directive is available. ++ ++The compiler will print these strings at the start of a new line, ++followed immediately by the object's initial value. In most cases, ++the string should contain a tab, a pseudo-op, and then another tab. ++ */ ++#undef TARGET_ASM_BYTE_OP ++#define TARGET_ASM_BYTE_OP "\t.byte\t" ++#undef TARGET_ASM_ALIGNED_HI_OP ++#define TARGET_ASM_ALIGNED_HI_OP "\t.align 1\n\t.short\t" ++#undef TARGET_ASM_ALIGNED_SI_OP ++#define TARGET_ASM_ALIGNED_SI_OP "\t.align 2\n\t.int\t" ++#undef TARGET_ASM_ALIGNED_DI_OP ++#define TARGET_ASM_ALIGNED_DI_OP NULL ++#undef TARGET_ASM_ALIGNED_TI_OP ++#define TARGET_ASM_ALIGNED_TI_OP NULL ++#undef TARGET_ASM_UNALIGNED_HI_OP ++#define TARGET_ASM_UNALIGNED_HI_OP "\t.short\t" ++#undef TARGET_ASM_UNALIGNED_SI_OP ++#define TARGET_ASM_UNALIGNED_SI_OP "\t.int\t" ++#undef TARGET_ASM_UNALIGNED_DI_OP ++#define TARGET_ASM_UNALIGNED_DI_OP NULL ++#undef TARGET_ASM_UNALIGNED_TI_OP ++#define TARGET_ASM_UNALIGNED_TI_OP NULL ++ ++#undef TARGET_ASM_OUTPUT_MI_THUNK ++#define TARGET_ASM_OUTPUT_MI_THUNK avr32_output_mi_thunk ++ ++#undef TARGET_ASM_CAN_OUTPUT_MI_THUNK ++#define TARGET_ASM_CAN_OUTPUT_MI_THUNK hook_bool_tree_hwi_hwi_tree_true ++ ++static void ++avr32_output_mi_thunk (FILE * file, ++ tree thunk ATTRIBUTE_UNUSED, ++ HOST_WIDE_INT delta, ++ HOST_WIDE_INT vcall_offset, tree function) ++ { ++ int mi_delta = delta; ++ int this_regno = ++ (aggregate_value_p (TREE_TYPE (TREE_TYPE (function)), function) ? ++ INTERNAL_REGNUM (11) : INTERNAL_REGNUM (12)); ++ ++ ++ if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") ++ || vcall_offset ++ || flag_pic) ++ { ++ fputs ("\tpushm\tlr\n", file); ++ } ++ ++ ++ if (mi_delta != 0) ++ { ++ if (avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21")) ++ { ++ fprintf (file, "\tsub\t%s, -0x%x\n", reg_names[this_regno], ++ mi_delta); ++ } ++ else ++ { ++ /* Immediate is larger than k21 we must make us a temp register by ++ pushing a register to the stack. */ ++ fprintf (file, "\tmov\tlr, lo(%x)\n", mi_delta); ++ fprintf (file, "\torh\tlr, hi(%x)\n", mi_delta); ++ fprintf (file, "\tadd\t%s, lr\n", reg_names[this_regno]); ++ } ++ } ++ ++ ++ if (vcall_offset != 0) ++ { ++ fprintf (file, "\tld.w\tlr, %s[0]\n", reg_names[this_regno]); ++ fprintf (file, "\tld.w\tlr, lr[%i]\n", (int) vcall_offset); ++ fprintf (file, "\tadd\t%s, lr\n", reg_names[this_regno]); ++ } ++ ++ ++ if ( (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") ++ || vcall_offset) ++ && !flag_pic ) ++ { ++ fputs ("\tpopm\tlr\n", file); ++ } ++ ++ if (flag_pic) ++ { ++ /* Load the got into lr and then load the pointer ++ to the function from the got and put it on the stack. ++ We can then call the function and restore lr by issuing ++ a doubleword load from the stack. We do not use a popm/ldm ++ since it will be treated as a return and might need a flushing ++ of the return-stack if available. */ ++ rtx label = gen_label_rtx (); ++ /* Load the got. */ ++ fputs ("\tlddpc\tlr, 0f\n", file); ++ (*targetm.asm_out.internal_label) (file, "L", ++ CODE_LABEL_NUMBER (label)); ++ fputs ("\trsub\tlr, pc\n", file); ++ /* Load the function pointer. */ ++ fputs ("\tld.w\tlr, lr[", file); ++ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); ++ fputs ("@got]\n", file); ++ /* Push the function pointer on the stack.*/ ++ fputs ("\tpushm\tlr\n", file); ++ /* Restore the old lr value and load the function pointer into ++ pc. */ ++ fputs ("\tld.d\tlr,sp++\n", file); ++ fprintf (file, "\t.align 2\n"); ++ fprintf (file, "0:\t.long\t.L%d - _GLOBAL_OFFSET_TABLE_\n", CODE_LABEL_NUMBER (label)); ++ } ++ else ++ { ++ fprintf (file, "\tlddpc\tpc, 0f\n"); ++ fprintf (file, "\t.align 2\n"); ++ fputs ("0:\t.long\t", file); ++ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); ++ fputc ('\n', file); ++ } ++ } ++ ++/* Implements target hook vector_mode_supported. */ ++bool ++avr32_vector_mode_supported (enum machine_mode mode) ++ { ++ if ((mode == V2HImode) || (mode == V4QImode)) ++ return true; ++ ++ return false; ++ } ++ ++ ++#undef TARGET_INIT_LIBFUNCS ++#define TARGET_INIT_LIBFUNCS avr32_init_libfuncs ++ ++#undef TARGET_INIT_BUILTINS ++#define TARGET_INIT_BUILTINS avr32_init_builtins ++ ++#undef TARGET_EXPAND_BUILTIN ++#define TARGET_EXPAND_BUILTIN avr32_expand_builtin ++ ++tree int_ftype_int, int_ftype_void, short_ftype_short, void_ftype_int_int, ++void_ftype_ptr_int; ++tree void_ftype_int, void_ftype_void, int_ftype_ptr_int; ++tree short_ftype_short, int_ftype_int_short, int_ftype_short_short, ++short_ftype_short_short; ++tree int_ftype_int_int, longlong_ftype_int_short, longlong_ftype_short_short; ++tree void_ftype_int_int_int_int_int, void_ftype_int_int_int; ++tree longlong_ftype_int_int, void_ftype_int_int_longlong; ++tree int_ftype_int_int_int, longlong_ftype_longlong_int_short; ++tree longlong_ftype_longlong_short_short, int_ftype_int_short_short; ++ ++#define def_builtin(NAME, TYPE, CODE) \ ++ lang_hooks.builtin_function ((NAME), (TYPE), (CODE), \ ++ BUILT_IN_MD, NULL, NULL_TREE) ++ ++#define def_mbuiltin(MASK, NAME, TYPE, CODE) \ ++ do \ ++ { \ ++ if ((MASK)) \ ++ lang_hooks.builtin_function ((NAME), (TYPE), (CODE), \ ++ BUILT_IN_MD, NULL, NULL_TREE); \ ++ } \ ++ while (0) ++ ++struct builtin_description ++{ ++ const unsigned int mask; ++ const enum insn_code icode; ++ const char *const name; ++ const int code; ++ const enum rtx_code comparison; ++ const unsigned int flag; ++ const tree *ftype; ++}; ++ ++static const struct builtin_description bdesc_2arg[] = { ++#define DSP_BUILTIN(code, builtin, ftype) \ ++ { 1, CODE_FOR_##code, "__builtin_" #code , \ ++ AVR32_BUILTIN_##builtin, 0, 0, ftype } ++ ++ DSP_BUILTIN (mulsathh_h, MULSATHH_H, &short_ftype_short_short), ++ DSP_BUILTIN (mulsathh_w, MULSATHH_W, &int_ftype_short_short), ++ DSP_BUILTIN (mulsatrndhh_h, MULSATRNDHH_H, &short_ftype_short_short), ++ DSP_BUILTIN (mulsatrndwh_w, MULSATRNDWH_W, &int_ftype_int_short), ++ DSP_BUILTIN (mulsatwh_w, MULSATWH_W, &int_ftype_int_short), ++ DSP_BUILTIN (satadd_h, SATADD_H, &short_ftype_short_short), ++ DSP_BUILTIN (satsub_h, SATSUB_H, &short_ftype_short_short), ++ DSP_BUILTIN (satadd_w, SATADD_W, &int_ftype_int_int), ++ DSP_BUILTIN (satsub_w, SATSUB_W, &int_ftype_int_int), ++ DSP_BUILTIN (mulwh_d, MULWH_D, &longlong_ftype_int_short), ++ DSP_BUILTIN (mulnwh_d, MULNWH_D, &longlong_ftype_int_short) ++}; ++ ++ ++void ++avr32_init_builtins (void) ++ { ++ unsigned int i; ++ const struct builtin_description *d; ++ tree endlink = void_list_node; ++ tree int_endlink = tree_cons (NULL_TREE, integer_type_node, endlink); ++ tree longlong_endlink = ++ tree_cons (NULL_TREE, long_long_integer_type_node, endlink); ++ tree short_endlink = ++ tree_cons (NULL_TREE, short_integer_type_node, endlink); ++ tree void_endlink = tree_cons (NULL_TREE, void_type_node, endlink); ++ ++ /* int func (int) */ ++ int_ftype_int = build_function_type (integer_type_node, int_endlink); ++ ++ /* short func (short) */ ++ short_ftype_short ++ = build_function_type (short_integer_type_node, short_endlink); ++ ++ /* short func (short, short) */ ++ short_ftype_short_short ++ = build_function_type (short_integer_type_node, ++ tree_cons (NULL_TREE, short_integer_type_node, ++ short_endlink)); ++ ++ /* long long func (long long, short, short) */ ++ longlong_ftype_longlong_short_short ++ = build_function_type (long_long_integer_type_node, ++ tree_cons (NULL_TREE, long_long_integer_type_node, ++ tree_cons (NULL_TREE, ++ short_integer_type_node, ++ short_endlink))); ++ ++ /* long long func (short, short) */ ++ longlong_ftype_short_short ++ = build_function_type (long_long_integer_type_node, ++ tree_cons (NULL_TREE, short_integer_type_node, ++ short_endlink)); ++ ++ /* int func (int, int) */ ++ int_ftype_int_int ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ int_endlink)); ++ ++ /* long long func (int, int) */ ++ longlong_ftype_int_int ++ = build_function_type (long_long_integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ int_endlink)); ++ ++ /* long long int func (long long, int, short) */ ++ longlong_ftype_longlong_int_short ++ = build_function_type (long_long_integer_type_node, ++ tree_cons (NULL_TREE, long_long_integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ short_endlink))); ++ ++ /* long long int func (int, short) */ ++ longlong_ftype_int_short ++ = build_function_type (long_long_integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ short_endlink)); ++ ++ /* int func (int, short, short) */ ++ int_ftype_int_short_short ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, ++ short_integer_type_node, ++ short_endlink))); ++ ++ /* int func (short, short) */ ++ int_ftype_short_short ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, short_integer_type_node, ++ short_endlink)); ++ ++ /* int func (int, short) */ ++ int_ftype_int_short ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ short_endlink)); ++ ++ /* void func (int, int) */ ++ void_ftype_int_int ++ = build_function_type (void_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ int_endlink)); ++ ++ /* void func (int, int, int) */ ++ void_ftype_int_int_int ++ = build_function_type (void_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ int_endlink))); ++ ++ /* void func (int, int, long long) */ ++ void_ftype_int_int_longlong ++ = build_function_type (void_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ longlong_endlink))); ++ ++ /* void func (int, int, int, int, int) */ ++ void_ftype_int_int_int_int_int ++ = build_function_type (void_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, ++ integer_type_node, ++ tree_cons ++ (NULL_TREE, ++ integer_type_node, ++ int_endlink))))); ++ ++ /* void func (void *, int) */ ++ void_ftype_ptr_int ++ = build_function_type (void_type_node, ++ tree_cons (NULL_TREE, ptr_type_node, int_endlink)); ++ ++ /* void func (int) */ ++ void_ftype_int = build_function_type (void_type_node, int_endlink); ++ ++ /* void func (void) */ ++ void_ftype_void = build_function_type (void_type_node, void_endlink); ++ ++ /* int func (void) */ ++ int_ftype_void = build_function_type (integer_type_node, void_endlink); ++ ++ /* int func (void *, int) */ ++ int_ftype_ptr_int ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, ptr_type_node, int_endlink)); ++ ++ /* int func (int, int, int) */ ++ int_ftype_int_int_int ++ = build_function_type (integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ tree_cons (NULL_TREE, integer_type_node, ++ int_endlink))); ++ ++ /* Initialize avr32 builtins. */ ++ def_builtin ("__builtin_mfsr", int_ftype_int, AVR32_BUILTIN_MFSR); ++ def_builtin ("__builtin_mtsr", void_ftype_int_int, AVR32_BUILTIN_MTSR); ++ def_builtin ("__builtin_mfdr", int_ftype_int, AVR32_BUILTIN_MFDR); ++ def_builtin ("__builtin_mtdr", void_ftype_int_int, AVR32_BUILTIN_MTDR); ++ def_builtin ("__builtin_cache", void_ftype_ptr_int, AVR32_BUILTIN_CACHE); ++ def_builtin ("__builtin_sync", void_ftype_int, AVR32_BUILTIN_SYNC); ++ def_builtin ("__builtin_tlbr", void_ftype_void, AVR32_BUILTIN_TLBR); ++ def_builtin ("__builtin_tlbs", void_ftype_void, AVR32_BUILTIN_TLBS); ++ def_builtin ("__builtin_tlbw", void_ftype_void, AVR32_BUILTIN_TLBW); ++ def_builtin ("__builtin_breakpoint", void_ftype_void, ++ AVR32_BUILTIN_BREAKPOINT); ++ def_builtin ("__builtin_xchg", int_ftype_ptr_int, AVR32_BUILTIN_XCHG); ++ def_builtin ("__builtin_ldxi", int_ftype_ptr_int, AVR32_BUILTIN_LDXI); ++ def_builtin ("__builtin_bswap_16", short_ftype_short, ++ AVR32_BUILTIN_BSWAP16); ++ def_builtin ("__builtin_bswap_32", int_ftype_int, AVR32_BUILTIN_BSWAP32); ++ def_builtin ("__builtin_cop", void_ftype_int_int_int_int_int, ++ AVR32_BUILTIN_COP); ++ def_builtin ("__builtin_mvcr_w", int_ftype_int_int, AVR32_BUILTIN_MVCR_W); ++ def_builtin ("__builtin_mvrc_w", void_ftype_int_int_int, ++ AVR32_BUILTIN_MVRC_W); ++ def_builtin ("__builtin_mvcr_d", longlong_ftype_int_int, ++ AVR32_BUILTIN_MVCR_D); ++ def_builtin ("__builtin_mvrc_d", void_ftype_int_int_longlong, ++ AVR32_BUILTIN_MVRC_D); ++ def_builtin ("__builtin_sats", int_ftype_int_int_int, AVR32_BUILTIN_SATS); ++ def_builtin ("__builtin_satu", int_ftype_int_int_int, AVR32_BUILTIN_SATU); ++ def_builtin ("__builtin_satrnds", int_ftype_int_int_int, ++ AVR32_BUILTIN_SATRNDS); ++ def_builtin ("__builtin_satrndu", int_ftype_int_int_int, ++ AVR32_BUILTIN_SATRNDU); ++ def_builtin ("__builtin_musfr", void_ftype_int, AVR32_BUILTIN_MUSFR); ++ def_builtin ("__builtin_mustr", int_ftype_void, AVR32_BUILTIN_MUSTR); ++ def_builtin ("__builtin_macsathh_w", int_ftype_int_short_short, ++ AVR32_BUILTIN_MACSATHH_W); ++ def_builtin ("__builtin_macwh_d", longlong_ftype_longlong_int_short, ++ AVR32_BUILTIN_MACWH_D); ++ def_builtin ("__builtin_machh_d", longlong_ftype_longlong_short_short, ++ AVR32_BUILTIN_MACHH_D); ++ ++ /* Add all builtins that are more or less simple operations on two ++ operands. */ ++ for (i = 0, d = bdesc_2arg; i < ARRAY_SIZE (bdesc_2arg); i++, d++) ++ { ++ /* Use one of the operands; the target can have a different mode for ++ mask-generating compares. */ ++ ++ if (d->name == 0) ++ continue; ++ ++ def_mbuiltin (d->mask, d->name, *(d->ftype), d->code); ++ } ++ } ++ ++ ++/* Subroutine of avr32_expand_builtin to take care of binop insns. */ ++ ++static rtx ++avr32_expand_binop_builtin (enum insn_code icode, tree arglist, rtx target) ++ { ++ rtx pat; ++ tree arg0 = TREE_VALUE (arglist); ++ tree arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ rtx op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ rtx op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ enum machine_mode tmode = insn_data[icode].operand[0].mode; ++ enum machine_mode mode0 = insn_data[icode].operand[1].mode; ++ enum machine_mode mode1 = insn_data[icode].operand[2].mode; ++ ++ if (!target ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ ++ /* In case the insn wants input operands in modes different from the ++ result, abort. */ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ /* If op0 is already a reg we must cast it to the correct mode. */ ++ if (REG_P (op0)) ++ op0 = convert_to_mode (mode0, op0, 1); ++ else ++ op0 = copy_to_mode_reg (mode0, op0); ++ } ++ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) ++ { ++ /* If op1 is already a reg we must cast it to the correct mode. */ ++ if (REG_P (op1)) ++ op1 = convert_to_mode (mode1, op1, 1); ++ else ++ op1 = copy_to_mode_reg (mode1, op1); ++ } ++ pat = GEN_FCN (icode) (target, op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ } ++ ++/* Expand an expression EXP that calls a built-in function, ++ with result going to TARGET if that's convenient ++ (and in mode MODE if that's convenient). ++ SUBTARGET may be used as the target for computing one of EXP's operands. ++ IGNORE is nonzero if the value is to be ignored. */ ++ ++rtx ++avr32_expand_builtin (tree exp, ++ rtx target, ++ rtx subtarget ATTRIBUTE_UNUSED, ++ enum machine_mode mode ATTRIBUTE_UNUSED, ++ int ignore ATTRIBUTE_UNUSED) ++ { ++ const struct builtin_description *d; ++ unsigned int i; ++ enum insn_code icode; ++ tree fndecl = TREE_OPERAND (TREE_OPERAND (exp, 0), 0); ++ tree arglist = TREE_OPERAND (exp, 1); ++ tree arg0, arg1, arg2; ++ rtx op0, op1, op2, pat; ++ enum machine_mode tmode, mode0, mode1; ++ enum machine_mode arg0_mode; ++ int fcode = DECL_FUNCTION_CODE (fndecl); ++ ++ switch (fcode) ++ { ++ default: ++ break; ++ ++ case AVR32_BUILTIN_SATS: ++ case AVR32_BUILTIN_SATU: ++ case AVR32_BUILTIN_SATRNDS: ++ case AVR32_BUILTIN_SATRNDU: ++ { ++ const char *fname; ++ switch (fcode) ++ { ++ default: ++ case AVR32_BUILTIN_SATS: ++ icode = CODE_FOR_sats; ++ fname = "sats"; ++ break; ++ case AVR32_BUILTIN_SATU: ++ icode = CODE_FOR_satu; ++ fname = "satu"; ++ break; ++ case AVR32_BUILTIN_SATRNDS: ++ icode = CODE_FOR_satrnds; ++ fname = "satrnds"; ++ break; ++ case AVR32_BUILTIN_SATRNDU: ++ icode = CODE_FOR_satrndu; ++ fname = "satrndu"; ++ break; ++ } ++ ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); ++ ++ tmode = insn_data[icode].operand[0].mode; ++ ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, GET_MODE (op0))) ++ { ++ op0 = copy_to_mode_reg (insn_data[icode].operand[0].mode, op0); ++ } ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) ++ { ++ error ("Parameter 2 to __builtin_%s should be a constant number.", ++ fname); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op2, SImode)) ++ { ++ error ("Parameter 3 to __builtin_%s should be a constant number.", ++ fname); ++ return NULL_RTX; ++ } ++ ++ emit_move_insn (target, op0); ++ pat = GEN_FCN (icode) (target, op1, op2); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return target; ++ } ++ case AVR32_BUILTIN_MUSTR: ++ icode = CODE_FOR_mustr; ++ tmode = insn_data[icode].operand[0].mode; ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ pat = GEN_FCN (icode) (target); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ ++ case AVR32_BUILTIN_MFSR: ++ icode = CODE_FOR_mfsr; ++ arg0 = TREE_VALUE (arglist); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ error ("Parameter 1 to __builtin_mfsr must be a constant number"); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ pat = GEN_FCN (icode) (target, op0); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ case AVR32_BUILTIN_MTSR: ++ icode = CODE_FOR_mtsr; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ mode0 = insn_data[icode].operand[0].mode; ++ mode1 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) ++ { ++ error ("Parameter 1 to __builtin_mtsr must be a constant number"); ++ return gen_reg_rtx (mode0); ++ } ++ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) ++ op1 = copy_to_mode_reg (mode1, op1); ++ pat = GEN_FCN (icode) (op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_MFDR: ++ icode = CODE_FOR_mfdr; ++ arg0 = TREE_VALUE (arglist); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ error ("Parameter 1 to __builtin_mfdr must be a constant number"); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ pat = GEN_FCN (icode) (target, op0); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ case AVR32_BUILTIN_MTDR: ++ icode = CODE_FOR_mtdr; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ mode0 = insn_data[icode].operand[0].mode; ++ mode1 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) ++ { ++ error ("Parameter 1 to __builtin_mtdr must be a constant number"); ++ return gen_reg_rtx (mode0); ++ } ++ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) ++ op1 = copy_to_mode_reg (mode1, op1); ++ pat = GEN_FCN (icode) (op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_CACHE: ++ icode = CODE_FOR_cache; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ mode0 = insn_data[icode].operand[0].mode; ++ mode1 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op1, mode1)) ++ { ++ error ("Parameter 2 to __builtin_cache must be a constant number"); ++ return gen_reg_rtx (mode1); ++ } ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) ++ op0 = copy_to_mode_reg (mode0, op0); ++ ++ pat = GEN_FCN (icode) (op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_SYNC: ++ case AVR32_BUILTIN_MUSFR: ++ { ++ const char *fname; ++ switch (fcode) ++ { ++ default: ++ case AVR32_BUILTIN_SYNC: ++ icode = CODE_FOR_sync; ++ fname = "sync"; ++ break; ++ case AVR32_BUILTIN_MUSFR: ++ icode = CODE_FOR_musfr; ++ fname = "musfr"; ++ break; ++ } ++ ++ arg0 = TREE_VALUE (arglist); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ mode0 = insn_data[icode].operand[0].mode; ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, mode0)) ++ { ++ if (icode == CODE_FOR_musfr) ++ op0 = copy_to_mode_reg (mode0, op0); ++ else ++ { ++ error ("Parameter to __builtin_%s is illegal.", fname); ++ return gen_reg_rtx (mode0); ++ } ++ } ++ pat = GEN_FCN (icode) (op0); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ } ++ case AVR32_BUILTIN_TLBR: ++ icode = CODE_FOR_tlbr; ++ pat = GEN_FCN (icode) (NULL_RTX); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_TLBS: ++ icode = CODE_FOR_tlbs; ++ pat = GEN_FCN (icode) (NULL_RTX); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_TLBW: ++ icode = CODE_FOR_tlbw; ++ pat = GEN_FCN (icode) (NULL_RTX); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_BREAKPOINT: ++ icode = CODE_FOR_breakpoint; ++ pat = GEN_FCN (icode) (NULL_RTX); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return NULL_RTX; ++ case AVR32_BUILTIN_XCHG: ++ icode = CODE_FOR_sync_lock_test_and_setsi; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ mode1 = insn_data[icode].operand[2].mode; ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) ++ { ++ op1 = copy_to_mode_reg (mode1, op1); ++ } ++ ++ op0 = gen_rtx_MEM (SImode, op0); ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ error ++ ("Parameter 1 to __builtin_xchg must be a pointer to an integer."); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ pat = GEN_FCN (icode) (target, op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ case AVR32_BUILTIN_LDXI: ++ icode = CODE_FOR_ldxi; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ mode1 = insn_data[icode].operand[2].mode; ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ op0 = copy_to_mode_reg (mode0, op0); ++ } ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op1, mode1)) ++ { ++ op1 = copy_to_mode_reg (mode1, op1); ++ } ++ ++ if (!(*insn_data[icode].operand[3].predicate) (op2, SImode)) ++ { ++ error ++ ("Parameter 3 to __builtin_ldxi must be a valid extract shift operand: (0|8|16|24)"); ++ return gen_reg_rtx (mode0); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ pat = GEN_FCN (icode) (target, op0, op1, op2); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ case AVR32_BUILTIN_BSWAP16: ++ { ++ icode = CODE_FOR_bswap_16; ++ arg0 = TREE_VALUE (arglist); ++ arg0_mode = TYPE_MODE (TREE_TYPE (arg0)); ++ mode0 = insn_data[icode].operand[1].mode; ++ if (arg0_mode != mode0) ++ arg0 = build1 (NOP_EXPR, ++ (*lang_hooks.types.type_for_mode) (mode0, 0), arg0); ++ ++ op0 = expand_expr (arg0, NULL_RTX, HImode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ op0 = copy_to_mode_reg (mode0, op0); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ { ++ target = gen_reg_rtx (tmode); ++ } ++ ++ ++ pat = GEN_FCN (icode) (target, op0); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return target; ++ } ++ case AVR32_BUILTIN_BSWAP32: ++ { ++ icode = CODE_FOR_bswap_32; ++ arg0 = TREE_VALUE (arglist); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) ++ { ++ op0 = copy_to_mode_reg (mode0, op0); ++ } ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ ++ ++ pat = GEN_FCN (icode) (target, op0); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return target; ++ } ++ case AVR32_BUILTIN_MVCR_W: ++ case AVR32_BUILTIN_MVCR_D: ++ { ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ ++ if (fcode == AVR32_BUILTIN_MVCR_W) ++ icode = CODE_FOR_mvcrsi; ++ else ++ icode = CODE_FOR_mvcrdi; ++ ++ tmode = insn_data[icode].operand[0].mode; ++ ++ if (target == 0 ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op0, SImode)) ++ { ++ error ++ ("Parameter 1 to __builtin_cop is not a valid coprocessor number."); ++ error ("Number should be between 0 and 7."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op1, SImode)) ++ { ++ error ++ ("Parameter 2 to __builtin_cop is not a valid coprocessor register number."); ++ error ("Number should be between 0 and 15."); ++ return NULL_RTX; ++ } ++ ++ pat = GEN_FCN (icode) (target, op0, op1); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return target; ++ } ++ case AVR32_BUILTIN_MACSATHH_W: ++ case AVR32_BUILTIN_MACWH_D: ++ case AVR32_BUILTIN_MACHH_D: ++ { ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); ++ ++ icode = ((fcode == AVR32_BUILTIN_MACSATHH_W) ? CODE_FOR_macsathh_w : ++ (fcode == AVR32_BUILTIN_MACWH_D) ? CODE_FOR_macwh_d : ++ CODE_FOR_machh_d); ++ ++ tmode = insn_data[icode].operand[0].mode; ++ mode0 = insn_data[icode].operand[1].mode; ++ mode1 = insn_data[icode].operand[2].mode; ++ ++ ++ if (!target ++ || GET_MODE (target) != tmode ++ || !(*insn_data[icode].operand[0].predicate) (target, tmode)) ++ target = gen_reg_rtx (tmode); ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, tmode)) ++ { ++ /* If op0 is already a reg we must cast it to the correct mode. */ ++ if (REG_P (op0)) ++ op0 = convert_to_mode (tmode, op0, 1); ++ else ++ op0 = copy_to_mode_reg (tmode, op0); ++ } ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op1, mode0)) ++ { ++ /* If op1 is already a reg we must cast it to the correct mode. */ ++ if (REG_P (op1)) ++ op1 = convert_to_mode (mode0, op1, 1); ++ else ++ op1 = copy_to_mode_reg (mode0, op1); ++ } ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op2, mode1)) ++ { ++ /* If op1 is already a reg we must cast it to the correct mode. */ ++ if (REG_P (op2)) ++ op2 = convert_to_mode (mode1, op2, 1); ++ else ++ op2 = copy_to_mode_reg (mode1, op2); ++ } ++ ++ emit_move_insn (target, op0); ++ ++ pat = GEN_FCN (icode) (target, op1, op2); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ return target; ++ } ++ case AVR32_BUILTIN_MVRC_W: ++ case AVR32_BUILTIN_MVRC_D: ++ { ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); ++ ++ if (fcode == AVR32_BUILTIN_MVRC_W) ++ icode = CODE_FOR_mvrcsi; ++ else ++ icode = CODE_FOR_mvrcdi; ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, SImode)) ++ { ++ error ("Parameter 1 is not a valid coprocessor number."); ++ error ("Number should be between 0 and 7."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) ++ { ++ error ("Parameter 2 is not a valid coprocessor register number."); ++ error ("Number should be between 0 and 15."); ++ return NULL_RTX; ++ } ++ ++ if (GET_CODE (op2) == CONST_INT ++ || GET_CODE (op2) == CONST ++ || GET_CODE (op2) == SYMBOL_REF || GET_CODE (op2) == LABEL_REF) ++ { ++ op2 = force_const_mem (insn_data[icode].operand[2].mode, op2); ++ } ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op2, GET_MODE (op2))) ++ op2 = copy_to_mode_reg (insn_data[icode].operand[2].mode, op2); ++ ++ ++ pat = GEN_FCN (icode) (op0, op1, op2); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return NULL_RTX; ++ } ++ case AVR32_BUILTIN_COP: ++ { ++ rtx op3, op4; ++ tree arg3, arg4; ++ icode = CODE_FOR_cop; ++ arg0 = TREE_VALUE (arglist); ++ arg1 = TREE_VALUE (TREE_CHAIN (arglist)); ++ arg2 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (arglist))); ++ arg3 = TREE_VALUE (TREE_CHAIN (TREE_CHAIN (TREE_CHAIN (arglist)))); ++ arg4 = ++ TREE_VALUE (TREE_CHAIN ++ (TREE_CHAIN (TREE_CHAIN (TREE_CHAIN (arglist))))); ++ op0 = expand_expr (arg0, NULL_RTX, VOIDmode, 0); ++ op1 = expand_expr (arg1, NULL_RTX, VOIDmode, 0); ++ op2 = expand_expr (arg2, NULL_RTX, VOIDmode, 0); ++ op3 = expand_expr (arg3, NULL_RTX, VOIDmode, 0); ++ op4 = expand_expr (arg4, NULL_RTX, VOIDmode, 0); ++ ++ if (!(*insn_data[icode].operand[0].predicate) (op0, SImode)) ++ { ++ error ++ ("Parameter 1 to __builtin_cop is not a valid coprocessor number."); ++ error ("Number should be between 0 and 7."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[1].predicate) (op1, SImode)) ++ { ++ error ++ ("Parameter 2 to __builtin_cop is not a valid coprocessor register number."); ++ error ("Number should be between 0 and 15."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[2].predicate) (op2, SImode)) ++ { ++ error ++ ("Parameter 3 to __builtin_cop is not a valid coprocessor register number."); ++ error ("Number should be between 0 and 15."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[3].predicate) (op3, SImode)) ++ { ++ error ++ ("Parameter 4 to __builtin_cop is not a valid coprocessor register number."); ++ error ("Number should be between 0 and 15."); ++ return NULL_RTX; ++ } ++ ++ if (!(*insn_data[icode].operand[4].predicate) (op4, SImode)) ++ { ++ error ++ ("Parameter 5 to __builtin_cop is not a valid coprocessor operation."); ++ error ("Number should be between 0 and 127."); ++ return NULL_RTX; ++ } ++ ++ pat = GEN_FCN (icode) (op0, op1, op2, op3, op4); ++ if (!pat) ++ return 0; ++ emit_insn (pat); ++ ++ return target; ++ } ++ } ++ ++ for (i = 0, d = bdesc_2arg; i < ARRAY_SIZE (bdesc_2arg); i++, d++) ++ if (d->code == fcode) ++ return avr32_expand_binop_builtin (d->icode, arglist, target); ++ ++ ++ /* @@@ Should really do something sensible here. */ ++ return NULL_RTX; ++ } ++ ++ ++/* Handle an "interrupt" or "isr" attribute; ++ arguments as in struct attribute_spec.handler. */ ++ ++static tree ++avr32_handle_isr_attribute (tree * node, tree name, tree args, ++ int flags, bool * no_add_attrs) ++ { ++ if (DECL_P (*node)) ++ { ++ if (TREE_CODE (*node) != FUNCTION_DECL) ++ { ++ warning ("`%s' attribute only applies to functions", ++ IDENTIFIER_POINTER (name)); ++ *no_add_attrs = true; ++ } ++ /* FIXME: the argument if any is checked for type attributes; should it ++ be checked for decl ones? */ ++ } ++ else ++ { ++ if (TREE_CODE (*node) == FUNCTION_TYPE ++ || TREE_CODE (*node) == METHOD_TYPE) ++ { ++ if (avr32_isr_value (args) == AVR32_FT_UNKNOWN) ++ { ++ warning ("`%s' attribute ignored", IDENTIFIER_POINTER (name)); ++ *no_add_attrs = true; ++ } ++ } ++ else if (TREE_CODE (*node) == POINTER_TYPE ++ && (TREE_CODE (TREE_TYPE (*node)) == FUNCTION_TYPE ++ || TREE_CODE (TREE_TYPE (*node)) == METHOD_TYPE) ++ && avr32_isr_value (args) != AVR32_FT_UNKNOWN) ++ { ++ *node = build_variant_type_copy (*node); ++ TREE_TYPE (*node) = build_type_attribute_variant ++ (TREE_TYPE (*node), ++ tree_cons (name, args, TYPE_ATTRIBUTES (TREE_TYPE (*node)))); ++ *no_add_attrs = true; ++ } ++ else ++ { ++ /* Possibly pass this attribute on from the type to a decl. */ ++ if (flags & ((int) ATTR_FLAG_DECL_NEXT ++ | (int) ATTR_FLAG_FUNCTION_NEXT ++ | (int) ATTR_FLAG_ARRAY_NEXT)) ++ { ++ *no_add_attrs = true; ++ return tree_cons (name, args, NULL_TREE); ++ } ++ else ++ { ++ warning ("`%s' attribute ignored", IDENTIFIER_POINTER (name)); ++ } ++ } ++ } ++ ++ return NULL_TREE; ++ } ++ ++/* Handle an attribute requiring a FUNCTION_DECL; ++ arguments as in struct attribute_spec.handler. */ ++static tree ++avr32_handle_fndecl_attribute (tree * node, tree name, ++ tree args ATTRIBUTE_UNUSED, ++ int flags ATTRIBUTE_UNUSED, ++ bool * no_add_attrs) ++ { ++ if (TREE_CODE (*node) != FUNCTION_DECL) ++ { ++ warning ("%qs attribute only applies to functions", ++ IDENTIFIER_POINTER (name)); ++ *no_add_attrs = true; ++ } ++ ++ return NULL_TREE; ++ } ++ ++ ++/* Handle an acall attribute; ++ arguments as in struct attribute_spec.handler. */ ++ ++static tree ++avr32_handle_acall_attribute (tree * node, tree name, ++ tree args ATTRIBUTE_UNUSED, ++ int flags ATTRIBUTE_UNUSED, bool * no_add_attrs) ++ { ++ if (TREE_CODE (*node) == FUNCTION_TYPE || TREE_CODE (*node) == METHOD_TYPE) ++ { ++ warning ("`%s' attribute not yet supported...", ++ IDENTIFIER_POINTER (name)); ++ *no_add_attrs = true; ++ return NULL_TREE; ++ } ++ ++ warning ("`%s' attribute only applies to functions", ++ IDENTIFIER_POINTER (name)); ++ *no_add_attrs = true; ++ return NULL_TREE; ++ } ++ ++ ++/* Return 0 if the attributes for two types are incompatible, 1 if they ++ are compatible, and 2 if they are nearly compatible (which causes a ++ warning to be generated). */ ++ ++static int ++avr32_comp_type_attributes (tree type1, tree type2) ++ { ++ int acall1, acall2, isr1, isr2, naked1, naked2; ++ ++ /* Check for mismatch of non-default calling convention. */ ++ if (TREE_CODE (type1) != FUNCTION_TYPE) ++ return 1; ++ ++ /* Check for mismatched call attributes. */ ++ acall1 = lookup_attribute ("acall", TYPE_ATTRIBUTES (type1)) != NULL; ++ acall2 = lookup_attribute ("acall", TYPE_ATTRIBUTES (type2)) != NULL; ++ naked1 = lookup_attribute ("naked", TYPE_ATTRIBUTES (type1)) != NULL; ++ naked2 = lookup_attribute ("naked", TYPE_ATTRIBUTES (type2)) != NULL; ++ isr1 = lookup_attribute ("isr", TYPE_ATTRIBUTES (type1)) != NULL; ++ if (!isr1) ++ isr1 = lookup_attribute ("interrupt", TYPE_ATTRIBUTES (type1)) != NULL; ++ ++ isr2 = lookup_attribute ("isr", TYPE_ATTRIBUTES (type2)) != NULL; ++ if (!isr2) ++ isr2 = lookup_attribute ("interrupt", TYPE_ATTRIBUTES (type2)) != NULL; ++ ++ if ((acall1 && isr2) ++ || (acall2 && isr1) || (naked1 && isr2) || (naked2 && isr1)) ++ return 0; ++ ++ return 1; ++ } ++ ++ ++/* Computes the type of the current function. */ ++ ++static unsigned long ++avr32_compute_func_type (void) ++ { ++ unsigned long type = AVR32_FT_UNKNOWN; ++ tree a; ++ tree attr; ++ ++ if (TREE_CODE (current_function_decl) != FUNCTION_DECL) ++ abort (); ++ ++ /* Decide if the current function is volatile. Such functions never ++ return, and many memory cycles can be saved by not storing register ++ values that will never be needed again. This optimization was added to ++ speed up context switching in a kernel application. */ ++ if (optimize > 0 ++ && TREE_NOTHROW (current_function_decl) ++ && TREE_THIS_VOLATILE (current_function_decl)) ++ type |= AVR32_FT_VOLATILE; ++ ++ if (cfun->static_chain_decl != NULL) ++ type |= AVR32_FT_NESTED; ++ ++ attr = DECL_ATTRIBUTES (current_function_decl); ++ ++ a = lookup_attribute ("isr", attr); ++ if (a == NULL_TREE) ++ a = lookup_attribute ("interrupt", attr); ++ ++ if (a == NULL_TREE) ++ type |= AVR32_FT_NORMAL; ++ else ++ type |= avr32_isr_value (TREE_VALUE (a)); ++ ++ ++ a = lookup_attribute ("acall", attr); ++ if (a != NULL_TREE) ++ type |= AVR32_FT_ACALL; ++ ++ a = lookup_attribute ("naked", attr); ++ if (a != NULL_TREE) ++ type |= AVR32_FT_NAKED; ++ ++ return type; ++ } ++ ++/* Returns the type of the current function. */ ++ ++static unsigned long ++avr32_current_func_type (void) ++ { ++ if (AVR32_FUNC_TYPE (cfun->machine->func_type) == AVR32_FT_UNKNOWN) ++ cfun->machine->func_type = avr32_compute_func_type (); ++ ++ return cfun->machine->func_type; ++ } ++ ++/* ++ This target hook should return true if we should not pass type solely ++ in registers. The file expr.h defines a definition that is usually appropriate, ++ refer to expr.h for additional documentation. ++ */ ++bool ++avr32_must_pass_in_stack (enum machine_mode mode ATTRIBUTE_UNUSED, tree type) ++ { ++ if (type && AGGREGATE_TYPE_P (type) ++ /* If the alignment is less than the size then pass in the struct on ++ the stack. */ ++ && ((unsigned int) TYPE_ALIGN_UNIT (type) < ++ (unsigned int) int_size_in_bytes (type)) ++ /* If we support unaligned word accesses then structs of size 4 and 8 ++ can have any alignment and still be passed in registers. */ ++ && !(TARGET_UNALIGNED_WORD ++ && (int_size_in_bytes (type) == 4 ++ || int_size_in_bytes (type) == 8)) ++ /* Double word structs need only a word alignment. */ ++ && !(int_size_in_bytes (type) == 8 && TYPE_ALIGN_UNIT (type) >= 4)) ++ return true; ++ ++ if (type && AGGREGATE_TYPE_P (type) ++ /* Structs of size 3,5,6,7 are always passed in registers. */ ++ && (int_size_in_bytes (type) == 3 ++ || int_size_in_bytes (type) == 5 ++ || int_size_in_bytes (type) == 6 || int_size_in_bytes (type) == 7)) ++ return true; ++ ++ ++ return (type && TREE_ADDRESSABLE (type)); ++ } ++ ++ ++bool ++avr32_strict_argument_naming (CUMULATIVE_ARGS * ca ATTRIBUTE_UNUSED) ++ { ++ return true; ++ } ++ ++/* ++ This target hook should return true if an argument at the position indicated ++ by cum should be passed by reference. This predicate is queried after target ++ independent reasons for being passed by reference, such as TREE_ADDRESSABLE (type). ++ ++ If the hook returns true, a copy of that argument is made in memory and a ++ pointer to the argument is passed instead of the argument itself. The pointer ++ is passed in whatever way is appropriate for passing a pointer to that type. ++ */ ++bool ++avr32_pass_by_reference (CUMULATIVE_ARGS * cum ATTRIBUTE_UNUSED, ++ enum machine_mode mode ATTRIBUTE_UNUSED, ++ tree type, bool named ATTRIBUTE_UNUSED) ++ { ++ return (type && (TREE_CODE (TYPE_SIZE (type)) != INTEGER_CST)); ++ } ++ ++static int ++avr32_arg_partial_bytes (CUMULATIVE_ARGS * pcum ATTRIBUTE_UNUSED, ++ enum machine_mode mode ATTRIBUTE_UNUSED, ++ tree type ATTRIBUTE_UNUSED, ++ bool named ATTRIBUTE_UNUSED) ++ { ++ return 0; ++ } ++ ++ ++struct gcc_target targetm = TARGET_INITIALIZER; ++ ++/* ++ Table used to convert from register number in the assembler instructions and ++ the register numbers used in gcc. ++ */ ++const int avr32_function_arg_reglist[] = { ++ INTERNAL_REGNUM (12), ++ INTERNAL_REGNUM (11), ++ INTERNAL_REGNUM (10), ++ INTERNAL_REGNUM (9), ++ INTERNAL_REGNUM (8) ++}; ++ ++rtx avr32_compare_op0 = NULL_RTX; ++rtx avr32_compare_op1 = NULL_RTX; ++rtx avr32_compare_operator = NULL_RTX; ++rtx avr32_acc_cache = NULL_RTX; ++ ++/* ++ Returns nonzero if it is allowed to store a value of mode mode in hard ++ register number regno. ++ */ ++int ++avr32_hard_regno_mode_ok (int regnr, enum machine_mode mode) ++ { ++ /* We allow only float modes in the fp-registers */ ++ if (regnr >= FIRST_FP_REGNUM ++ && regnr <= LAST_FP_REGNUM && GET_MODE_CLASS (mode) != MODE_FLOAT) ++ { ++ return 0; ++ } ++ ++ switch (mode) ++ { ++ case DImode: /* long long */ ++ case DFmode: /* double */ ++ case SCmode: /* __complex__ float */ ++ case CSImode: /* __complex__ int */ ++ if (regnr < 4) ++ { /* long long int not supported in r12, sp, lr ++ or pc. */ ++ return 0; ++ } ++ else ++ { ++ if (regnr % 2) /* long long int has to be refered in even ++ registers. */ ++ return 0; ++ else ++ return 1; ++ } ++ case CDImode: /* __complex__ long long */ ++ case DCmode: /* __complex__ double */ ++ case TImode: /* 16 bytes */ ++ if (regnr < 7) ++ return 0; ++ else if (regnr % 2) ++ return 0; ++ else ++ return 1; ++ default: ++ return 1; ++ } ++ } ++ ++ ++int ++avr32_rnd_operands (rtx add, rtx shift) ++ { ++ if (GET_CODE (shift) == CONST_INT && ++ GET_CODE (add) == CONST_INT && INTVAL (shift) > 0) ++ { ++ if ((1 << (INTVAL (shift) - 1)) == INTVAL (add)) ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++ ++ ++int ++avr32_const_ok_for_constraint_p (HOST_WIDE_INT value, char c, const char *str) ++ { ++ switch (c) ++ { ++ case 'K': ++ case 'I': ++ { ++ HOST_WIDE_INT min_value = 0, max_value = 0; ++ char size_str[3]; ++ int const_size; ++ ++ size_str[0] = str[2]; ++ size_str[1] = str[3]; ++ size_str[2] = '\0'; ++ const_size = atoi (size_str); ++ ++ if (toupper (str[1]) == 'U') ++ { ++ min_value = 0; ++ max_value = (1 << const_size) - 1; ++ } ++ else if (toupper (str[1]) == 'S') ++ { ++ min_value = -(1 << (const_size - 1)); ++ max_value = (1 << (const_size - 1)) - 1; ++ } ++ ++ if (c == 'I') ++ { ++ value = -value; ++ } ++ ++ if (value >= min_value && value <= max_value) ++ { ++ return 1; ++ } ++ break; ++ } ++ case 'M': ++ return avr32_mask_upper_bits_operand (GEN_INT (value), VOIDmode); ++ } ++ ++ return 0; ++ } ++ ++ ++/*Compute mask of which floating-point registers needs saving upon ++ entry to this function*/ ++static unsigned long ++avr32_compute_save_fp_reg_mask (void) ++ { ++ unsigned long func_type = avr32_current_func_type (); ++ unsigned int save_reg_mask = 0; ++ unsigned int reg; ++ unsigned int max_reg = 7; ++ int save_all_call_used_regs = FALSE; ++ ++ /* This only applies for hardware floating-point implementation. */ ++ if (!TARGET_HARD_FLOAT) ++ return 0; ++ ++ if (IS_INTERRUPT (func_type)) ++ { ++ ++ /* Interrupt functions must not corrupt any registers, even call ++ clobbered ones. If this is a leaf function we can just examine the ++ registers used by the RTL, but otherwise we have to assume that ++ whatever function is called might clobber anything, and so we have ++ to save all the call-clobbered registers as well. */ ++ max_reg = 13; ++ save_all_call_used_regs = !current_function_is_leaf; ++ } ++ ++ /* All used registers used must be saved */ ++ for (reg = 0; reg <= max_reg; reg++) ++ if (regs_ever_live[INTERNAL_FP_REGNUM (reg)] ++ || (save_all_call_used_regs ++ && call_used_regs[INTERNAL_FP_REGNUM (reg)])) ++ save_reg_mask |= (1 << reg); ++ ++ return save_reg_mask; ++ } ++ ++/*Compute mask of registers which needs saving upon function entry */ ++static unsigned long ++avr32_compute_save_reg_mask (int push) ++ { ++ unsigned long func_type; ++ unsigned int save_reg_mask = 0; ++ unsigned int reg; ++ ++ func_type = avr32_current_func_type (); ++ ++ if (IS_INTERRUPT (func_type)) ++ { ++ unsigned int max_reg = 12; ++ ++ ++ /* Get the banking scheme for the interrupt */ ++ switch (func_type) ++ { ++ case AVR32_FT_ISR_FULL: ++ max_reg = 0; ++ break; ++ case AVR32_FT_ISR_HALF: ++ max_reg = 7; ++ break; ++ case AVR32_FT_ISR_NONE: ++ max_reg = 12; ++ break; ++ } ++ ++ /* Interrupt functions must not corrupt any registers, even call ++ clobbered ones. If this is a leaf function we can just examine the ++ registers used by the RTL, but otherwise we have to assume that ++ whatever function is called might clobber anything, and so we have ++ to save all the call-clobbered registers as well. */ ++ ++ /* Need not push the registers r8-r12 for AVR32A architectures, as this ++ is automatially done in hardware. We also do not have any shadow ++ registers. */ ++ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) ++ { ++ max_reg = 7; ++ func_type = AVR32_FT_ISR_NONE; ++ } ++ ++ /* All registers which are used and is not shadowed must be saved */ ++ for (reg = 0; reg <= max_reg; reg++) ++ if (regs_ever_live[INTERNAL_REGNUM (reg)] ++ || (!current_function_is_leaf ++ && call_used_regs[INTERNAL_REGNUM (reg)])) ++ save_reg_mask |= (1 << reg); ++ ++ /* Check LR */ ++ if ((regs_ever_live[LR_REGNUM] ++ || !current_function_is_leaf || frame_pointer_needed) ++ /* Only non-shadowed register models */ ++ && (func_type == AVR32_FT_ISR_NONE)) ++ save_reg_mask |= (1 << ASM_REGNUM (LR_REGNUM)); ++ ++ /* Make sure that the GOT register is pushed. */ ++ if (max_reg >= ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM) ++ && current_function_uses_pic_offset_table) ++ save_reg_mask |= (1 << ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM)); ++ ++ } ++ else ++ { ++ int use_pushm = optimize_size; ++ ++ /* In the normal case we only need to save those registers which are ++ call saved and which are used by this function. */ ++ for (reg = 0; reg <= 7; reg++) ++ if (regs_ever_live[INTERNAL_REGNUM (reg)] ++ && !call_used_regs[INTERNAL_REGNUM (reg)]) ++ save_reg_mask |= (1 << reg); ++ ++ /* Make sure that the GOT register is pushed. */ ++ if (current_function_uses_pic_offset_table) ++ save_reg_mask |= (1 << ASM_REGNUM (PIC_OFFSET_TABLE_REGNUM)); ++ ++ ++ /* If we optimize for size and do not have anonymous arguments: use ++ popm/pushm always */ ++ if (use_pushm) ++ { ++ if ((save_reg_mask & (1 << 0)) ++ || (save_reg_mask & (1 << 1)) ++ || (save_reg_mask & (1 << 2)) || (save_reg_mask & (1 << 3))) ++ save_reg_mask |= 0xf; ++ ++ if ((save_reg_mask & (1 << 4)) ++ || (save_reg_mask & (1 << 5)) ++ || (save_reg_mask & (1 << 6)) || (save_reg_mask & (1 << 7))) ++ save_reg_mask |= 0xf0; ++ ++ if ((save_reg_mask & (1 << 8)) || (save_reg_mask & (1 << 9))) ++ save_reg_mask |= 0x300; ++ } ++ ++ ++ /* Check LR */ ++ if ((regs_ever_live[LR_REGNUM] ++ || !current_function_is_leaf ++ || (optimize_size ++ && save_reg_mask ++ && !current_function_calls_eh_return) || frame_pointer_needed)) ++ { ++ if (push ++ /* Never pop LR into PC for functions which ++ calls __builtin_eh_return, since we need to ++ fix the SP after the restoring of the registers ++ and before returning. */ ++ || current_function_calls_eh_return) ++ { ++ /* Push/Pop LR */ ++ save_reg_mask |= (1 << ASM_REGNUM (LR_REGNUM)); ++ } ++ else ++ { ++ /* Pop PC */ ++ save_reg_mask |= (1 << ASM_REGNUM (PC_REGNUM)); ++ } ++ } ++ } ++ ++ ++ /* Save registers so the exception handler can modify them. */ ++ if (current_function_calls_eh_return) ++ { ++ unsigned int i; ++ ++ for (i = 0;; i++) ++ { ++ reg = EH_RETURN_DATA_REGNO (i); ++ if (reg == INVALID_REGNUM) ++ break; ++ save_reg_mask |= 1 << ASM_REGNUM (reg); ++ } ++ } ++ ++ return save_reg_mask; ++ } ++ ++/*Compute total size in bytes of all saved registers */ ++static int ++avr32_get_reg_mask_size (int reg_mask) ++ { ++ int reg, size; ++ size = 0; ++ ++ for (reg = 0; reg <= 15; reg++) ++ if (reg_mask & (1 << reg)) ++ size += 4; ++ ++ return size; ++ } ++ ++/*Get a register from one of the registers which are saved onto the stack ++ upon function entry */ ++ ++static int ++avr32_get_saved_reg (int save_reg_mask) ++ { ++ unsigned int reg; ++ ++ /* Find the first register which is saved in the saved_reg_mask */ ++ for (reg = 0; reg <= 15; reg++) ++ if (save_reg_mask & (1 << reg)) ++ return reg; ++ ++ return -1; ++ } ++ ++/* Return 1 if it is possible to return using a single instruction. */ ++int ++avr32_use_return_insn (int iscond) ++ { ++ unsigned int func_type = avr32_current_func_type (); ++ unsigned long saved_int_regs; ++ unsigned long saved_fp_regs; ++ ++ /* Never use a return instruction before reload has run. */ ++ if (!reload_completed) ++ return 0; ++ ++ /* Must adjust the stack for vararg functions. */ ++ if (current_function_args_info.uses_anonymous_args) ++ return 0; ++ ++ /* If there a stack adjstment. */ ++ if (get_frame_size ()) ++ return 0; ++ ++ saved_int_regs = avr32_compute_save_reg_mask (TRUE); ++ saved_fp_regs = avr32_compute_save_fp_reg_mask (); ++ ++ /* Functions which have saved fp-regs on the stack can not be performed in ++ one instruction */ ++ if (saved_fp_regs) ++ return 0; ++ ++ /* Conditional returns can not be performed in one instruction if we need ++ to restore registers from the stack */ ++ if (iscond && saved_int_regs) ++ return 0; ++ ++ /* Conditional return can not be used for interrupt handlers. */ ++ if (iscond && IS_INTERRUPT (func_type)) ++ return 0; ++ ++ /* For interrupt handlers which needs to pop registers */ ++ if (saved_int_regs && IS_INTERRUPT (func_type)) ++ return 0; ++ ++ ++ /* If there are saved registers but the LR isn't saved, then we need two ++ instructions for the return. */ ++ if (saved_int_regs && !(saved_int_regs & (1 << ASM_REGNUM (LR_REGNUM)))) ++ return 0; ++ ++ ++ return 1; ++ } ++ ++ ++/*Generate some function prologue info in the assembly file*/ ++ ++void ++avr32_target_asm_function_prologue (FILE * f, HOST_WIDE_INT frame_size) ++ { ++ if (IS_NAKED (avr32_current_func_type ())) ++ fprintf (f, ++ "\t# Function is naked: Prologue and epilogue provided by programmer\n"); ++ ++ if (IS_INTERRUPT (avr32_current_func_type ())) ++ { ++ switch (avr32_current_func_type ()) ++ { ++ case AVR32_FT_ISR_FULL: ++ fprintf (f, ++ "\t# Interrupt Function: Fully shadowed register file\n"); ++ break; ++ case AVR32_FT_ISR_HALF: ++ fprintf (f, ++ "\t# Interrupt Function: Half shadowed register file\n"); ++ break; ++ default: ++ case AVR32_FT_ISR_NONE: ++ fprintf (f, "\t# Interrupt Function: No shadowed register file\n"); ++ break; ++ } ++ } ++ ++ ++ fprintf (f, "\t# args = %i, frame = %li, pretend = %i\n", ++ current_function_args_size, frame_size, ++ current_function_pretend_args_size); ++ ++ fprintf (f, "\t# frame_needed = %i, leaf_function = %i\n", ++ frame_pointer_needed, current_function_is_leaf); ++ ++ fprintf (f, "\t# uses_anonymous_args = %i\n", ++ current_function_args_info.uses_anonymous_args); ++ if (current_function_calls_eh_return) ++ fprintf (f, "\t# Calls __builtin_eh_return.\n"); ++ ++ } ++ ++ ++/* Generate and emit an insn that we will recognize as a pushm or stm. ++ Unfortunately, since this insn does not reflect very well the actual ++ semantics of the operation, we need to annotate the insn for the benefit ++ of DWARF2 frame unwind information. */ ++ ++int avr32_convert_to_reglist16 (int reglist8_vect); ++ ++static rtx ++emit_multi_reg_push (int reglist, int usePUSHM) ++ { ++ rtx insn; ++ rtx dwarf; ++ rtx tmp; ++ rtx reg; ++ int i; ++ int nr_regs; ++ int index = 0; ++ ++ if (usePUSHM) ++ { ++ insn = emit_insn (gen_pushm (gen_rtx_CONST_INT (SImode, reglist))); ++ reglist = avr32_convert_to_reglist16 (reglist); ++ } ++ else ++ { ++ insn = emit_insn (gen_stm (stack_pointer_rtx, ++ gen_rtx_CONST_INT (SImode, reglist), ++ gen_rtx_CONST_INT (SImode, 1))); ++ } ++ ++ nr_regs = avr32_get_reg_mask_size (reglist) / 4; ++ dwarf = gen_rtx_SEQUENCE (VOIDmode, rtvec_alloc (nr_regs + 1)); ++ ++ for (i = 15; i >= 0; i--) ++ { ++ if (reglist & (1 << i)) ++ { ++ reg = gen_rtx_REG (SImode, INTERNAL_REGNUM (i)); ++ tmp = gen_rtx_SET (VOIDmode, ++ gen_rtx_MEM (SImode, ++ plus_constant (stack_pointer_rtx, ++ 4 * index)), reg); ++ RTX_FRAME_RELATED_P (tmp) = 1; ++ XVECEXP (dwarf, 0, 1 + index++) = tmp; ++ } ++ } ++ ++ tmp = gen_rtx_SET (SImode, ++ stack_pointer_rtx, ++ gen_rtx_PLUS (SImode, ++ stack_pointer_rtx, ++ GEN_INT (-4 * nr_regs))); ++ RTX_FRAME_RELATED_P (tmp) = 1; ++ XVECEXP (dwarf, 0, 0) = tmp; ++ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, dwarf, ++ REG_NOTES (insn)); ++ return insn; ++ } ++ ++ ++static rtx ++emit_multi_fp_reg_push (int reglist) ++ { ++ rtx insn; ++ rtx dwarf; ++ rtx tmp; ++ rtx reg; ++ int i; ++ int nr_regs; ++ int index = 0; ++ ++ insn = emit_insn (gen_stm_fp (stack_pointer_rtx, ++ gen_rtx_CONST_INT (SImode, reglist), ++ gen_rtx_CONST_INT (SImode, 1))); ++ ++ nr_regs = avr32_get_reg_mask_size (reglist) / 4; ++ dwarf = gen_rtx_SEQUENCE (VOIDmode, rtvec_alloc (nr_regs + 1)); ++ ++ for (i = 15; i >= 0; i--) ++ { ++ if (reglist & (1 << i)) ++ { ++ reg = gen_rtx_REG (SImode, INTERNAL_FP_REGNUM (i)); ++ tmp = gen_rtx_SET (VOIDmode, ++ gen_rtx_MEM (SImode, ++ plus_constant (stack_pointer_rtx, ++ 4 * index)), reg); ++ RTX_FRAME_RELATED_P (tmp) = 1; ++ XVECEXP (dwarf, 0, 1 + index++) = tmp; ++ } ++ } ++ ++ tmp = gen_rtx_SET (SImode, ++ stack_pointer_rtx, ++ gen_rtx_PLUS (SImode, ++ stack_pointer_rtx, ++ GEN_INT (-4 * nr_regs))); ++ RTX_FRAME_RELATED_P (tmp) = 1; ++ XVECEXP (dwarf, 0, 0) = tmp; ++ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, dwarf, ++ REG_NOTES (insn)); ++ return insn; ++ } ++ ++rtx ++avr32_gen_load_multiple (rtx * regs, int count, rtx from, ++ int write_back, int in_struct_p, int scalar_p) ++ { ++ ++ rtx result; ++ int i = 0, j; ++ ++ result = ++ gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (count + (write_back ? 1 : 0))); ++ ++ if (write_back) ++ { ++ XVECEXP (result, 0, 0) ++ = gen_rtx_SET (GET_MODE (from), from, ++ plus_constant (from, count * 4)); ++ i = 1; ++ count++; ++ } ++ ++ ++ for (j = 0; i < count; i++, j++) ++ { ++ rtx unspec; ++ rtx mem = gen_rtx_MEM (SImode, plus_constant (from, j * 4)); ++ MEM_IN_STRUCT_P (mem) = in_struct_p; ++ MEM_SCALAR_P (mem) = scalar_p; ++ unspec = gen_rtx_UNSPEC (VOIDmode, gen_rtvec (1, mem), UNSPEC_LDM); ++ XVECEXP (result, 0, i) = gen_rtx_SET (VOIDmode, regs[j], unspec); ++ } ++ ++ return result; ++ } ++ ++ ++rtx ++avr32_gen_store_multiple (rtx * regs, int count, rtx to, ++ int in_struct_p, int scalar_p) ++ { ++ rtx result; ++ int i = 0, j; ++ ++ result = gen_rtx_PARALLEL (VOIDmode, rtvec_alloc (count)); ++ ++ for (j = 0; i < count; i++, j++) ++ { ++ rtx mem = gen_rtx_MEM (SImode, plus_constant (to, j * 4)); ++ MEM_IN_STRUCT_P (mem) = in_struct_p; ++ MEM_SCALAR_P (mem) = scalar_p; ++ XVECEXP (result, 0, i) ++ = gen_rtx_SET (VOIDmode, mem, ++ gen_rtx_UNSPEC (VOIDmode, ++ gen_rtvec (1, regs[j]), ++ UNSPEC_STORE_MULTIPLE)); ++ } ++ ++ return result; ++ } ++ ++ ++/* Move a block of memory if it is word aligned or we support unaligned ++ word memory accesses. The size must be maximum 64 bytes. */ ++ ++int ++avr32_gen_movmemsi (rtx * operands) ++ { ++ HOST_WIDE_INT bytes_to_go; ++ rtx src, dst; ++ rtx st_src, st_dst; ++ int ptr_offset = 0; ++ int block_size; ++ int dst_in_struct_p, src_in_struct_p; ++ int dst_scalar_p, src_scalar_p; ++ int unaligned; ++ ++ if (GET_CODE (operands[2]) != CONST_INT ++ || GET_CODE (operands[3]) != CONST_INT ++ || INTVAL (operands[2]) > 64 ++ || ((INTVAL (operands[3]) & 3) && !TARGET_UNALIGNED_WORD)) ++ return 0; ++ ++ unaligned = (INTVAL (operands[3]) & 3) != 0; ++ ++ block_size = 4; ++ ++ st_dst = XEXP (operands[0], 0); ++ st_src = XEXP (operands[1], 0); ++ ++ dst_in_struct_p = MEM_IN_STRUCT_P (operands[0]); ++ dst_scalar_p = MEM_SCALAR_P (operands[0]); ++ src_in_struct_p = MEM_IN_STRUCT_P (operands[1]); ++ src_scalar_p = MEM_SCALAR_P (operands[1]); ++ ++ dst = copy_to_mode_reg (SImode, st_dst); ++ src = copy_to_mode_reg (SImode, st_src); ++ ++ bytes_to_go = INTVAL (operands[2]); ++ ++ while (bytes_to_go) ++ { ++ enum machine_mode move_mode; ++ /* (Seems to be a problem with reloads for the movti pattern so this is ++ disabled until that problem is resolved) ++ UPDATE: Problem seems to be solved now.... */ ++ if (bytes_to_go >= GET_MODE_SIZE (TImode) && !unaligned ++ /* Do not emit ldm/stm for UC3 as ld.d/st.d is more optimal. */ ++ && avr32_arch->arch_type != ARCH_TYPE_AVR32_UC) ++ move_mode = TImode; ++ else if ((bytes_to_go >= GET_MODE_SIZE (DImode)) && !unaligned) ++ move_mode = DImode; ++ else if (bytes_to_go >= GET_MODE_SIZE (SImode)) ++ move_mode = SImode; ++ else ++ move_mode = QImode; ++ ++ { ++ rtx dst_mem = gen_rtx_MEM (move_mode, ++ gen_rtx_PLUS (SImode, dst, ++ GEN_INT (ptr_offset))); ++ rtx src_mem = gen_rtx_MEM (move_mode, ++ gen_rtx_PLUS (SImode, src, ++ GEN_INT (ptr_offset))); ++ ptr_offset += GET_MODE_SIZE (move_mode); ++ bytes_to_go -= GET_MODE_SIZE (move_mode); ++ ++ MEM_IN_STRUCT_P (dst_mem) = dst_in_struct_p; ++ MEM_SCALAR_P (dst_mem) = dst_scalar_p; ++ ++ MEM_IN_STRUCT_P (src_mem) = src_in_struct_p; ++ MEM_SCALAR_P (src_mem) = src_scalar_p; ++ emit_move_insn (dst_mem, src_mem); ++ ++ } ++ } ++ ++ return 1; ++ } ++ ++ ++ ++/*Expand the prologue instruction*/ ++void ++avr32_expand_prologue (void) ++ { ++ rtx insn, dwarf; ++ unsigned long saved_reg_mask, saved_fp_reg_mask; ++ int reglist8 = 0; ++ ++ /* Naked functions does not have a prologue */ ++ if (IS_NAKED (avr32_current_func_type ())) ++ return; ++ ++ saved_reg_mask = avr32_compute_save_reg_mask (TRUE); ++ ++ if (saved_reg_mask) ++ { ++ /* Must push used registers */ ++ ++ /* Should we use POPM or LDM? */ ++ int usePUSHM = TRUE; ++ reglist8 = 0; ++ if (((saved_reg_mask & (1 << 0)) || ++ (saved_reg_mask & (1 << 1)) || ++ (saved_reg_mask & (1 << 2)) || (saved_reg_mask & (1 << 3)))) ++ { ++ /* One of R0-R3 should at least be pushed */ ++ if (((saved_reg_mask & (1 << 0)) && ++ (saved_reg_mask & (1 << 1)) && ++ (saved_reg_mask & (1 << 2)) && (saved_reg_mask & (1 << 3)))) ++ { ++ /* All should be pushed */ ++ reglist8 |= 0x01; ++ } ++ else ++ { ++ usePUSHM = FALSE; ++ } ++ } ++ ++ if (((saved_reg_mask & (1 << 4)) || ++ (saved_reg_mask & (1 << 5)) || ++ (saved_reg_mask & (1 << 6)) || (saved_reg_mask & (1 << 7)))) ++ { ++ /* One of R4-R7 should at least be pushed */ ++ if (((saved_reg_mask & (1 << 4)) && ++ (saved_reg_mask & (1 << 5)) && ++ (saved_reg_mask & (1 << 6)) && (saved_reg_mask & (1 << 7)))) ++ { ++ if (usePUSHM) ++ /* All should be pushed */ ++ reglist8 |= 0x02; ++ } ++ else ++ { ++ usePUSHM = FALSE; ++ } ++ } ++ ++ if (((saved_reg_mask & (1 << 8)) || (saved_reg_mask & (1 << 9)))) ++ { ++ /* One of R8-R9 should at least be pushed */ ++ if (((saved_reg_mask & (1 << 8)) && (saved_reg_mask & (1 << 9)))) ++ { ++ if (usePUSHM) ++ /* All should be pushed */ ++ reglist8 |= 0x04; ++ } ++ else ++ { ++ usePUSHM = FALSE; ++ } ++ } ++ ++ if (saved_reg_mask & (1 << 10)) ++ reglist8 |= 0x08; ++ ++ if (saved_reg_mask & (1 << 11)) ++ reglist8 |= 0x10; ++ ++ if (saved_reg_mask & (1 << 12)) ++ reglist8 |= 0x20; ++ ++ if (saved_reg_mask & (1 << ASM_REGNUM (LR_REGNUM))) ++ { ++ /* Push LR */ ++ reglist8 |= 0x40; ++ } ++ ++ if (usePUSHM) ++ { ++ insn = emit_multi_reg_push (reglist8, TRUE); ++ } ++ else ++ { ++ insn = emit_multi_reg_push (saved_reg_mask, FALSE); ++ } ++ RTX_FRAME_RELATED_P (insn) = 1; ++ ++ /* Prevent this instruction from being scheduled after any other ++ instructions. */ ++ emit_insn (gen_blockage ()); ++ } ++ ++ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); ++ if (saved_fp_reg_mask) ++ { ++ insn = emit_multi_fp_reg_push (saved_fp_reg_mask); ++ RTX_FRAME_RELATED_P (insn) = 1; ++ ++ /* Prevent this instruction from being scheduled after any other ++ instructions. */ ++ emit_insn (gen_blockage ()); ++ } ++ ++ /* Set frame pointer */ ++ if (frame_pointer_needed) ++ { ++ insn = emit_move_insn (frame_pointer_rtx, stack_pointer_rtx); ++ RTX_FRAME_RELATED_P (insn) = 1; ++ } ++ ++ if (get_frame_size () > 0) ++ { ++ if (avr32_const_ok_for_constraint_p (get_frame_size (), 'K', "Ks21")) ++ { ++ insn = emit_insn (gen_rtx_SET (SImode, ++ stack_pointer_rtx, ++ gen_rtx_PLUS (SImode, ++ stack_pointer_rtx, ++ gen_rtx_CONST_INT ++ (SImode, ++ -get_frame_size ++ ())))); ++ RTX_FRAME_RELATED_P (insn) = 1; ++ } ++ else ++ { ++ /* Immediate is larger than k21 We must either check if we can use ++ one of the pushed reegisters as temporary storage or we must ++ make us a temp register by pushing a register to the stack. */ ++ rtx temp_reg, const_pool_entry, insn; ++ if (saved_reg_mask) ++ { ++ temp_reg = ++ gen_rtx_REG (SImode, ++ INTERNAL_REGNUM (avr32_get_saved_reg ++ (saved_reg_mask))); ++ } ++ else ++ { ++ temp_reg = gen_rtx_REG (SImode, INTERNAL_REGNUM (7)); ++ emit_move_insn (gen_rtx_MEM ++ (SImode, ++ gen_rtx_PRE_DEC (SImode, stack_pointer_rtx)), ++ temp_reg); ++ } ++ ++ const_pool_entry = ++ force_const_mem (SImode, ++ gen_rtx_CONST_INT (SImode, get_frame_size ())); ++ emit_move_insn (temp_reg, const_pool_entry); ++ ++ insn = emit_insn (gen_rtx_SET (SImode, ++ stack_pointer_rtx, ++ gen_rtx_MINUS (SImode, ++ stack_pointer_rtx, ++ temp_reg))); ++ ++ dwarf = gen_rtx_SET (VOIDmode, stack_pointer_rtx, ++ gen_rtx_PLUS (SImode, stack_pointer_rtx, ++ GEN_INT (-get_frame_size ()))); ++ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_FRAME_RELATED_EXPR, ++ dwarf, REG_NOTES (insn)); ++ RTX_FRAME_RELATED_P (insn) = 1; ++ ++ if (!saved_reg_mask) ++ { ++ insn = ++ emit_move_insn (temp_reg, ++ gen_rtx_MEM (SImode, ++ gen_rtx_POST_INC (SImode, ++ gen_rtx_REG ++ (SImode, ++ 13)))); ++ } ++ ++ /* Mark the temp register as dead */ ++ REG_NOTES (insn) = gen_rtx_EXPR_LIST (REG_DEAD, temp_reg, ++ REG_NOTES (insn)); ++ ++ ++ } ++ ++ /* Prevent the the stack adjustment to be scheduled after any ++ instructions using the frame pointer. */ ++ emit_insn (gen_blockage ()); ++ } ++ ++ /* Load GOT */ ++ if (flag_pic) ++ { ++ avr32_load_pic_register (); ++ ++ /* gcc does not know that load or call instructions might use the pic ++ register so it might schedule these instructions before the loading ++ of the pic register. To avoid this emit a barrier for now. TODO! ++ Find out a better way to let gcc know which instructions might use ++ the pic register. */ ++ emit_insn (gen_blockage ()); ++ } ++ return; ++ } ++ ++void ++avr32_set_return_address (rtx source, rtx scratch) ++ { ++ rtx addr; ++ unsigned long saved_regs; ++ ++ saved_regs = avr32_compute_save_reg_mask (TRUE); ++ ++ if (!(saved_regs & (1 << ASM_REGNUM (LR_REGNUM)))) ++ emit_move_insn (gen_rtx_REG (Pmode, LR_REGNUM), source); ++ else ++ { ++ if (frame_pointer_needed) ++ addr = gen_rtx_REG (Pmode, FRAME_POINTER_REGNUM); ++ else ++ if (avr32_const_ok_for_constraint_p (get_frame_size (), 'K', "Ks16")) ++ { ++ addr = plus_constant (stack_pointer_rtx, get_frame_size ()); ++ } ++ else ++ { ++ emit_insn (gen_movsi (scratch, GEN_INT (get_frame_size ()))); ++ addr = scratch; ++ } ++ emit_move_insn (gen_rtx_MEM (Pmode, addr), source); ++ } ++ } ++ ++ ++ ++/* Return the length of INSN. LENGTH is the initial length computed by ++ attributes in the machine-description file. */ ++ ++int ++avr32_adjust_insn_length (rtx insn ATTRIBUTE_UNUSED, ++ int length ATTRIBUTE_UNUSED) ++ { ++ return length; ++ } ++ ++void ++avr32_output_return_instruction (int single_ret_inst ATTRIBUTE_UNUSED, ++ int iscond ATTRIBUTE_UNUSED, ++ rtx cond ATTRIBUTE_UNUSED, rtx r12_imm) ++ { ++ ++ unsigned long saved_reg_mask, saved_fp_reg_mask; ++ int insert_ret = TRUE; ++ int reglist8 = 0; ++ int stack_adjustment = get_frame_size (); ++ unsigned int func_type = avr32_current_func_type (); ++ FILE *f = asm_out_file; ++ ++ /* Naked functions does not have an epilogue */ ++ if (IS_NAKED (func_type)) ++ return; ++ ++ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); ++ ++ saved_reg_mask = avr32_compute_save_reg_mask (FALSE); ++ ++ /* Reset frame pointer */ ++ if (stack_adjustment > 0) ++ { ++ if (avr32_const_ok_for_constraint_p (stack_adjustment, 'I', "Is21")) ++ { ++ fprintf (f, "\tsub\tsp, %i # Reset Frame Pointer\n", ++ -stack_adjustment); ++ } ++ else ++ { ++ /* TODO! Is it safe to use r8 as scratch?? */ ++ fprintf (f, "\tmov\tr8, lo(%i) # Reset Frame Pointer\n", ++ -stack_adjustment); ++ fprintf (f, "\torh\tr8, hi(%i) # Reset Frame Pointer\n", ++ -stack_adjustment); ++ fprintf (f, "\tadd\tsp, r8 # Reset Frame Pointer\n"); ++ } ++ } ++ ++ if (saved_fp_reg_mask) ++ { ++ char reglist[64]; /* 64 bytes should be enough... */ ++ avr32_make_fp_reglist_w (saved_fp_reg_mask, (char *) reglist); ++ fprintf (f, "\tldcm.w\tcp0, sp++, %s\n", reglist); ++ if (saved_fp_reg_mask & ~0xff) ++ { ++ saved_fp_reg_mask &= ~0xff; ++ avr32_make_fp_reglist_d (saved_fp_reg_mask, (char *) reglist); ++ fprintf (f, "\tldcm.d\tcp0, sp++, %s\n", reglist); ++ } ++ } ++ ++ if (saved_reg_mask) ++ { ++ /* Must pop used registers */ ++ ++ /* Should we use POPM or LDM? */ ++ int usePOPM = TRUE; ++ if (((saved_reg_mask & (1 << 0)) || ++ (saved_reg_mask & (1 << 1)) || ++ (saved_reg_mask & (1 << 2)) || (saved_reg_mask & (1 << 3)))) ++ { ++ /* One of R0-R3 should at least be popped */ ++ if (((saved_reg_mask & (1 << 0)) && ++ (saved_reg_mask & (1 << 1)) && ++ (saved_reg_mask & (1 << 2)) && (saved_reg_mask & (1 << 3)))) ++ { ++ /* All should be popped */ ++ reglist8 |= 0x01; ++ } ++ else ++ { ++ usePOPM = FALSE; ++ } ++ } ++ ++ if (((saved_reg_mask & (1 << 4)) || ++ (saved_reg_mask & (1 << 5)) || ++ (saved_reg_mask & (1 << 6)) || (saved_reg_mask & (1 << 7)))) ++ { ++ /* One of R0-R3 should at least be popped */ ++ if (((saved_reg_mask & (1 << 4)) && ++ (saved_reg_mask & (1 << 5)) && ++ (saved_reg_mask & (1 << 6)) && (saved_reg_mask & (1 << 7)))) ++ { ++ if (usePOPM) ++ /* All should be popped */ ++ reglist8 |= 0x02; ++ } ++ else ++ { ++ usePOPM = FALSE; ++ } ++ } ++ ++ if (((saved_reg_mask & (1 << 8)) || (saved_reg_mask & (1 << 9)))) ++ { ++ /* One of R8-R9 should at least be pushed */ ++ if (((saved_reg_mask & (1 << 8)) && (saved_reg_mask & (1 << 9)))) ++ { ++ if (usePOPM) ++ /* All should be pushed */ ++ reglist8 |= 0x04; ++ } ++ else ++ { ++ usePOPM = FALSE; ++ } ++ } ++ ++ if (saved_reg_mask & (1 << 10)) ++ reglist8 |= 0x08; ++ ++ if (saved_reg_mask & (1 << 11)) ++ reglist8 |= 0x10; ++ ++ if (saved_reg_mask & (1 << 12)) ++ reglist8 |= 0x20; ++ ++ if (saved_reg_mask & (1 << ASM_REGNUM (LR_REGNUM))) ++ /* Pop LR */ ++ reglist8 |= 0x40; ++ ++ if (saved_reg_mask & (1 << ASM_REGNUM (PC_REGNUM))) ++ /* Pop LR into PC. */ ++ reglist8 |= 0x80; ++ ++ if (usePOPM) ++ { ++ char reglist[64]; /* 64 bytes should be enough... */ ++ avr32_make_reglist8 (reglist8, (char *) reglist); ++ ++ if (reglist8 & 0x80) ++ /* This instruction is also a return */ ++ insert_ret = FALSE; ++ ++ if (r12_imm && !insert_ret) ++ fprintf (f, "\tpopm\t%s, r12=%li\n", reglist, INTVAL (r12_imm)); ++ else ++ fprintf (f, "\tpopm\t%s\n", reglist); ++ ++ } ++ else ++ { ++ char reglist[64]; /* 64 bytes should be enough... */ ++ avr32_make_reglist16 (saved_reg_mask, (char *) reglist); ++ if (saved_reg_mask & (1 << ASM_REGNUM (PC_REGNUM))) ++ /* This instruction is also a return */ ++ insert_ret = FALSE; ++ ++ if (r12_imm && !insert_ret) ++ fprintf (f, "\tldm\tsp++, %s, r12=%li\n", reglist, ++ INTVAL (r12_imm)); ++ else ++ fprintf (f, "\tldm\tsp++, %s\n", reglist); ++ ++ } ++ ++ } ++ ++ /* Stack adjustment for exception handler. */ ++ if (current_function_calls_eh_return) ++ fprintf (f, "\tadd\tsp, r%d\n", ASM_REGNUM (EH_RETURN_STACKADJ_REGNO)); ++ ++ ++ if (IS_INTERRUPT (func_type)) ++ { ++ fprintf (f, "\trete\n"); ++ } ++ else if (insert_ret) ++ { ++ if (r12_imm) ++ fprintf (f, "\tretal\t%li\n", INTVAL (r12_imm)); ++ else ++ fprintf (f, "\tretal\tr12\n"); ++ } ++ } ++ ++/* Function for converting a fp-register mask to a ++ reglistCPD8 register list string. */ ++void ++avr32_make_fp_reglist_d (int reglist_mask, char *reglist_string) ++ { ++ int i; ++ ++ /* Make sure reglist_string is empty */ ++ reglist_string[0] = '\0'; ++ ++ for (i = 0; i < NUM_FP_REGS; i += 2) ++ { ++ if (reglist_mask & (1 << i)) ++ { ++ strlen (reglist_string) ? ++ sprintf (reglist_string, "%s, %s-%s", reglist_string, ++ reg_names[INTERNAL_FP_REGNUM (i)], ++ reg_names[INTERNAL_FP_REGNUM (i + 1)]) : ++ sprintf (reglist_string, "%s-%s", ++ reg_names[INTERNAL_FP_REGNUM (i)], ++ reg_names[INTERNAL_FP_REGNUM (i + 1)]); ++ } ++ } ++ } ++ ++/* Function for converting a fp-register mask to a ++ reglistCP8 register list string. */ ++void ++avr32_make_fp_reglist_w (int reglist_mask, char *reglist_string) ++ { ++ int i; ++ ++ /* Make sure reglist_string is empty */ ++ reglist_string[0] = '\0'; ++ ++ for (i = 0; i < NUM_FP_REGS; ++i) ++ { ++ if (reglist_mask & (1 << i)) ++ { ++ strlen (reglist_string) ? ++ sprintf (reglist_string, "%s, %s", reglist_string, ++ reg_names[INTERNAL_FP_REGNUM (i)]) : ++ sprintf (reglist_string, "%s", reg_names[INTERNAL_FP_REGNUM (i)]); ++ } ++ } ++ } ++ ++void ++avr32_make_reglist16 (int reglist16_vect, char *reglist16_string) ++ { ++ int i; ++ ++ /* Make sure reglist16_string is empty */ ++ reglist16_string[0] = '\0'; ++ ++ for (i = 0; i < 16; ++i) ++ { ++ if (reglist16_vect & (1 << i)) ++ { ++ strlen (reglist16_string) ? ++ sprintf (reglist16_string, "%s, %s", reglist16_string, ++ reg_names[INTERNAL_REGNUM (i)]) : ++ sprintf (reglist16_string, "%s", reg_names[INTERNAL_REGNUM (i)]); ++ } ++ } ++ } ++ ++int ++avr32_convert_to_reglist16 (int reglist8_vect) ++ { ++ int reglist16_vect = 0; ++ if (reglist8_vect & 0x1) ++ reglist16_vect |= 0xF; ++ if (reglist8_vect & 0x2) ++ reglist16_vect |= 0xF0; ++ if (reglist8_vect & 0x4) ++ reglist16_vect |= 0x300; ++ if (reglist8_vect & 0x8) ++ reglist16_vect |= 0x400; ++ if (reglist8_vect & 0x10) ++ reglist16_vect |= 0x800; ++ if (reglist8_vect & 0x20) ++ reglist16_vect |= 0x1000; ++ if (reglist8_vect & 0x40) ++ reglist16_vect |= 0x4000; ++ if (reglist8_vect & 0x80) ++ reglist16_vect |= 0x8000; ++ ++ return reglist16_vect; ++ } ++ ++void ++avr32_make_reglist8 (int reglist8_vect, char *reglist8_string) ++ { ++ /* Make sure reglist8_string is empty */ ++ reglist8_string[0] = '\0'; ++ ++ if (reglist8_vect & 0x1) ++ sprintf (reglist8_string, "r0-r3"); ++ if (reglist8_vect & 0x2) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r4-r7", ++ reglist8_string) : ++ sprintf (reglist8_string, "r4-r7"); ++ if (reglist8_vect & 0x4) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r8-r9", ++ reglist8_string) : ++ sprintf (reglist8_string, "r8-r9"); ++ if (reglist8_vect & 0x8) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r10", ++ reglist8_string) : ++ sprintf (reglist8_string, "r10"); ++ if (reglist8_vect & 0x10) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r11", ++ reglist8_string) : ++ sprintf (reglist8_string, "r11"); ++ if (reglist8_vect & 0x20) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, r12", ++ reglist8_string) : ++ sprintf (reglist8_string, "r12"); ++ if (reglist8_vect & 0x40) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, lr", ++ reglist8_string) : ++ sprintf (reglist8_string, "lr"); ++ if (reglist8_vect & 0x80) ++ strlen (reglist8_string) ? sprintf (reglist8_string, "%s, pc", ++ reglist8_string) : ++ sprintf (reglist8_string, "pc"); ++ } ++ ++int ++avr32_eh_return_data_regno (int n) ++ { ++ if (n >= 0 && n <= 3) ++ return 8 + n; ++ else ++ return INVALID_REGNUM; ++ } ++ ++/* Compute the distance from register FROM to register TO. ++ These can be the arg pointer, the frame pointer or ++ the stack pointer. ++ Typical stack layout looks like this: ++ ++ old stack pointer -> | | ++ ---- ++ | | \ ++ | | saved arguments for ++ | | vararg functions ++ arg_pointer -> | | / ++ -- ++ | | \ ++ | | call saved ++ | | registers ++ | | / ++ frame ptr -> -- ++ | | \ ++ | | local ++ | | variables ++ stack ptr --> | | / ++ -- ++ | | \ ++ | | outgoing ++ | | arguments ++ | | / ++ -- ++ ++ For a given funciton some or all of these stack compomnents ++ may not be needed, giving rise to the possibility of ++ eliminating some of the registers. ++ ++ The values returned by this function must reflect the behaviour ++ of avr32_expand_prologue() and avr32_compute_save_reg_mask(). ++ ++ The sign of the number returned reflects the direction of stack ++ growth, so the values are positive for all eliminations except ++ from the soft frame pointer to the hard frame pointer. */ ++ ++ ++int ++avr32_initial_elimination_offset (int from, int to) ++ { ++ int i; ++ int call_saved_regs = 0; ++ unsigned long saved_reg_mask, saved_fp_reg_mask; ++ unsigned int local_vars = get_frame_size (); ++ ++ saved_reg_mask = avr32_compute_save_reg_mask (TRUE); ++ saved_fp_reg_mask = avr32_compute_save_fp_reg_mask (); ++ ++ for (i = 0; i < 16; ++i) ++ { ++ if (saved_reg_mask & (1 << i)) ++ call_saved_regs += 4; ++ } ++ ++ for (i = 0; i < NUM_FP_REGS; ++i) ++ { ++ if (saved_fp_reg_mask & (1 << i)) ++ call_saved_regs += 4; ++ } ++ ++ switch (from) ++ { ++ case ARG_POINTER_REGNUM: ++ switch (to) ++ { ++ case STACK_POINTER_REGNUM: ++ return call_saved_regs + local_vars; ++ case FRAME_POINTER_REGNUM: ++ return call_saved_regs; ++ default: ++ abort (); ++ } ++ case FRAME_POINTER_REGNUM: ++ switch (to) ++ { ++ case STACK_POINTER_REGNUM: ++ return local_vars; ++ default: ++ abort (); ++ } ++ default: ++ abort (); ++ } ++ } ++ ++ ++/* ++ Returns a rtx used when passing the next argument to a function. ++ avr32_init_cumulative_args() and avr32_function_arg_advance() sets witch ++ register to use. ++ */ ++rtx ++avr32_function_arg (CUMULATIVE_ARGS * cum, enum machine_mode mode, ++ tree type, int named) ++ { ++ int index = -1; ++ ++ HOST_WIDE_INT arg_size, arg_rsize; ++ if (type) ++ { ++ arg_size = int_size_in_bytes (type); ++ } ++ else ++ { ++ arg_size = GET_MODE_SIZE (mode); ++ } ++ arg_rsize = PUSH_ROUNDING (arg_size); ++ ++ /* ++ The last time this macro is called, it is called with mode == VOIDmode, ++ and its result is passed to the call or call_value pattern as operands 2 ++ and 3 respectively. */ ++ if (mode == VOIDmode) ++ { ++ return gen_rtx_CONST_INT (SImode, 22); /* ToDo: fixme. */ ++ } ++ ++ if ((*targetm.calls.must_pass_in_stack) (mode, type) || !named) ++ { ++ return NULL_RTX; ++ } ++ ++ if (arg_rsize == 8) ++ { ++ /* use r11:r10 or r9:r8. */ ++ if (!(GET_USED_INDEX (cum, 1) || GET_USED_INDEX (cum, 2))) ++ index = 1; ++ else if (!(GET_USED_INDEX (cum, 3) || GET_USED_INDEX (cum, 4))) ++ index = 3; ++ else ++ index = -1; ++ } ++ else if (arg_rsize == 4) ++ { /* Use first available register */ ++ index = 0; ++ while (index <= LAST_CUM_REG_INDEX && GET_USED_INDEX (cum, index)) ++ index++; ++ if (index > LAST_CUM_REG_INDEX) ++ index = -1; ++ } ++ ++ SET_REG_INDEX (cum, index); ++ ++ if (GET_REG_INDEX (cum) >= 0) ++ return gen_rtx_REG (mode, ++ avr32_function_arg_reglist[GET_REG_INDEX (cum)]); ++ ++ return NULL_RTX; ++ } ++ ++/* ++ Set the register used for passing the first argument to a function. ++ */ ++void ++avr32_init_cumulative_args (CUMULATIVE_ARGS * cum, tree fntype, ++ rtx libname ATTRIBUTE_UNUSED, ++ tree fndecl ATTRIBUTE_UNUSED) ++ { ++ /* Set all registers as unused. */ ++ SET_INDEXES_UNUSED (cum); ++ ++ /* Reset uses_anonymous_args */ ++ cum->uses_anonymous_args = 0; ++ ++ /* Reset size of stack pushed arguments */ ++ cum->stack_pushed_args_size = 0; ++ } ++ ++/* ++ Set register used for passing the next argument to a function. Only the ++ Scratch Registers are used. ++ ++ number name ++ 15 r15 PC ++ 14 r14 LR ++ 13 r13 _SP_________ ++ FIRST_CUM_REG 12 r12 _||_ ++ 10 r11 || ++ 11 r10 _||_ Scratch Registers ++ 8 r9 || ++ LAST_SCRATCH_REG 9 r8 _\/_________ ++ 6 r7 /\ ++ 7 r6 || ++ 4 r5 || ++ 5 r4 || ++ 2 r3 || ++ 3 r2 || ++ 0 r1 || ++ 1 r0 _||_________ ++ ++ */ ++void ++avr32_function_arg_advance (CUMULATIVE_ARGS * cum, enum machine_mode mode, ++ tree type, int named ATTRIBUTE_UNUSED) ++ { ++ HOST_WIDE_INT arg_size, arg_rsize; ++ ++ if (type) ++ { ++ arg_size = int_size_in_bytes (type); ++ } ++ else ++ { ++ arg_size = GET_MODE_SIZE (mode); ++ } ++ arg_rsize = PUSH_ROUNDING (arg_size); ++ ++ /* It the argument had to be passed in stack, no register is used. */ ++ if ((*targetm.calls.must_pass_in_stack) (mode, type)) ++ { ++ cum->stack_pushed_args_size += PUSH_ROUNDING (int_size_in_bytes (type)); ++ return; ++ } ++ ++ /* Mark the used registers as "used". */ ++ if (GET_REG_INDEX (cum) >= 0) ++ { ++ SET_USED_INDEX (cum, GET_REG_INDEX (cum)); ++ if (arg_rsize == 8) ++ { ++ SET_USED_INDEX (cum, (GET_REG_INDEX (cum) + 1)); ++ } ++ } ++ else ++ { ++ /* Had to use stack */ ++ cum->stack_pushed_args_size += arg_rsize; ++ } ++ } ++ ++/* ++ Defines witch direction to go to find the next register to use if the ++ argument is larger then one register or for arguments shorter than an ++ int which is not promoted, such as the last part of structures with ++ size not a multiple of 4. */ ++enum direction ++avr32_function_arg_padding (enum machine_mode mode ATTRIBUTE_UNUSED, ++ tree type) ++ { ++ /* Pad upward for all aggregates except byte and halfword sized aggregates ++ which can be passed in registers. */ ++ if (type ++ && AGGREGATE_TYPE_P (type) ++ && (int_size_in_bytes (type) != 1) ++ && !((int_size_in_bytes (type) == 2) ++ && TYPE_ALIGN_UNIT (type) >= 2) ++ && (int_size_in_bytes (type) & 0x3)) ++ { ++ return upward; ++ } ++ ++ return downward; ++ } ++ ++/* ++ Return a rtx used for the return value from a function call. ++ */ ++rtx ++avr32_function_value (tree type, tree func, bool outgoing ATTRIBUTE_UNUSED) ++ { ++ if (avr32_return_in_memory (type, func)) ++ return NULL_RTX; ++ ++ if (int_size_in_bytes (type) <= 4) ++ if (avr32_return_in_msb (type)) ++ /* Aggregates of size less than a word which does align the data in the ++ MSB must use SImode for r12. */ ++ return gen_rtx_REG (SImode, RET_REGISTER); ++ else ++ return gen_rtx_REG (TYPE_MODE (type), RET_REGISTER); ++ else if (int_size_in_bytes (type) <= 8) ++ return gen_rtx_REG (TYPE_MODE (type), INTERNAL_REGNUM (11)); ++ ++ return NULL_RTX; ++ } ++ ++/* ++ Return a rtx used for the return value from a library function call. ++ */ ++rtx ++avr32_libcall_value (enum machine_mode mode) ++ { ++ ++ if (GET_MODE_SIZE (mode) <= 4) ++ return gen_rtx_REG (mode, RET_REGISTER); ++ else if (GET_MODE_SIZE (mode) <= 8) ++ return gen_rtx_REG (mode, INTERNAL_REGNUM (11)); ++ else ++ return NULL_RTX; ++ } ++ ++/* Return TRUE if X references a SYMBOL_REF. */ ++int ++symbol_mentioned_p (rtx x) ++ { ++ const char *fmt; ++ int i; ++ ++ if (GET_CODE (x) == SYMBOL_REF) ++ return 1; ++ ++ fmt = GET_RTX_FORMAT (GET_CODE (x)); ++ ++ for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--) ++ { ++ if (fmt[i] == 'E') ++ { ++ int j; ++ ++ for (j = XVECLEN (x, i) - 1; j >= 0; j--) ++ if (symbol_mentioned_p (XVECEXP (x, i, j))) ++ return 1; ++ } ++ else if (fmt[i] == 'e' && symbol_mentioned_p (XEXP (x, i))) ++ return 1; ++ } ++ ++ return 0; ++ } ++ ++/* Return TRUE if X references a LABEL_REF. */ ++int ++label_mentioned_p (rtx x) ++ { ++ const char *fmt; ++ int i; ++ ++ if (GET_CODE (x) == LABEL_REF) ++ return 1; ++ ++ fmt = GET_RTX_FORMAT (GET_CODE (x)); ++ for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--) ++ { ++ if (fmt[i] == 'E') ++ { ++ int j; ++ ++ for (j = XVECLEN (x, i) - 1; j >= 0; j--) ++ if (label_mentioned_p (XVECEXP (x, i, j))) ++ return 1; ++ } ++ else if (fmt[i] == 'e' && label_mentioned_p (XEXP (x, i))) ++ return 1; ++ } ++ ++ return 0; ++ } ++ ++ ++int ++avr32_legitimate_pic_operand_p (rtx x) ++ { ++ ++ /* We can't have const, this must be broken down to a symbol. */ ++ if (GET_CODE (x) == CONST) ++ return FALSE; ++ ++ /* Can't access symbols or labels via the constant pool either */ ++ if ((GET_CODE (x) == SYMBOL_REF ++ && CONSTANT_POOL_ADDRESS_P (x) ++ && (symbol_mentioned_p (get_pool_constant (x)) ++ || label_mentioned_p (get_pool_constant (x))))) ++ return FALSE; ++ ++ return TRUE; ++ } ++ ++ ++rtx ++legitimize_pic_address (rtx orig, enum machine_mode mode ATTRIBUTE_UNUSED, ++ rtx reg) ++ { ++ ++ if (GET_CODE (orig) == SYMBOL_REF || GET_CODE (orig) == LABEL_REF) ++ { ++ int subregs = 0; ++ ++ if (reg == 0) ++ { ++ if (no_new_pseudos) ++ abort (); ++ else ++ reg = gen_reg_rtx (Pmode); ++ ++ subregs = 1; ++ } ++ ++ emit_move_insn (reg, orig); ++ ++ /* Only set current function as using pic offset table if flag_pic is ++ set. This is because this function is also used if ++ TARGET_HAS_ASM_ADDR_PSEUDOS is set. */ ++ if (flag_pic) ++ current_function_uses_pic_offset_table = 1; ++ ++ /* Put a REG_EQUAL note on this insn, so that it can be optimized by ++ loop. */ ++ return reg; ++ } ++ else if (GET_CODE (orig) == CONST) ++ { ++ rtx base, offset; ++ ++ if (flag_pic ++ && GET_CODE (XEXP (orig, 0)) == PLUS ++ && XEXP (XEXP (orig, 0), 0) == pic_offset_table_rtx) ++ return orig; ++ ++ if (reg == 0) ++ { ++ if (no_new_pseudos) ++ abort (); ++ else ++ reg = gen_reg_rtx (Pmode); ++ } ++ ++ if (GET_CODE (XEXP (orig, 0)) == PLUS) ++ { ++ base = ++ legitimize_pic_address (XEXP (XEXP (orig, 0), 0), Pmode, reg); ++ offset = ++ legitimize_pic_address (XEXP (XEXP (orig, 0), 1), Pmode, ++ base == reg ? 0 : reg); ++ } ++ else ++ abort (); ++ ++ if (GET_CODE (offset) == CONST_INT) ++ { ++ /* The base register doesn't really matter, we only want to test ++ the index for the appropriate mode. */ ++ if (!avr32_const_ok_for_constraint_p (INTVAL (offset), 'I', "Is21")) ++ { ++ if (!no_new_pseudos) ++ offset = force_reg (Pmode, offset); ++ else ++ abort (); ++ } ++ ++ if (GET_CODE (offset) == CONST_INT) ++ return plus_constant (base, INTVAL (offset)); ++ } ++ ++ return gen_rtx_PLUS (Pmode, base, offset); ++ } ++ ++ return orig; ++ } ++ ++/* Generate code to load the PIC register. */ ++void ++avr32_load_pic_register (void) ++ { ++ rtx l1, pic_tmp; ++ rtx global_offset_table; ++ ++ if ((current_function_uses_pic_offset_table == 0) || TARGET_NO_INIT_GOT) ++ return; ++ ++ if (!flag_pic) ++ abort (); ++ ++ l1 = gen_label_rtx (); ++ ++ global_offset_table = gen_rtx_SYMBOL_REF (Pmode, "_GLOBAL_OFFSET_TABLE_"); ++ pic_tmp = ++ gen_rtx_CONST (Pmode, ++ gen_rtx_MINUS (SImode, gen_rtx_LABEL_REF (Pmode, l1), ++ global_offset_table)); ++ emit_insn (gen_pic_load_addr ++ (pic_offset_table_rtx, force_const_mem (SImode, pic_tmp))); ++ emit_insn (gen_pic_compute_got_from_pc (pic_offset_table_rtx, l1)); ++ ++ /* Need to emit this whether or not we obey regdecls, since setjmp/longjmp ++ can cause life info to screw up. */ ++ emit_insn (gen_rtx_USE (VOIDmode, pic_offset_table_rtx)); ++ } ++ ++ ++ ++/* This hook should return true if values of type type are returned at the most ++ significant end of a register (in other words, if they are padded at the ++ least significant end). You can assume that type is returned in a register; ++ the caller is required to check this. Note that the register provided by ++ FUNCTION_VALUE must be able to hold the complete return value. For example, ++ if a 1-, 2- or 3-byte structure is returned at the most significant end of a ++ 4-byte register, FUNCTION_VALUE should provide an SImode rtx. */ ++bool ++avr32_return_in_msb (tree type ATTRIBUTE_UNUSED) ++ { ++ /* if ( AGGREGATE_TYPE_P (type) ) if ((int_size_in_bytes(type) == 1) || ++ ((int_size_in_bytes(type) == 2) && TYPE_ALIGN_UNIT(type) >= 2)) return ++ false; else return true; */ ++ ++ return false; ++ } ++ ++ ++/* ++ Returns one if a certain function value is going to be returned in memory ++ and zero if it is going to be returned in a register. ++ ++ BLKmode and all other modes that is larger than 64 bits are returned in ++ memory. ++ */ ++bool ++avr32_return_in_memory (tree type, tree fntype ATTRIBUTE_UNUSED) ++ { ++ if (TYPE_MODE (type) == VOIDmode) ++ return false; ++ ++ if (int_size_in_bytes (type) > (2 * UNITS_PER_WORD) ++ || int_size_in_bytes (type) == -1) ++ { ++ return true; ++ } ++ ++ /* If we have an aggregate then use the same mechanism as when checking if ++ it should be passed on the stack. */ ++ if (type ++ && AGGREGATE_TYPE_P (type) ++ && (*targetm.calls.must_pass_in_stack) (TYPE_MODE (type), type)) ++ return true; ++ ++ return false; ++ } ++ ++ ++/* Output the constant part of the trampoline. ++ lddpc r0, pc[0x8:e] ; load static chain register ++ lddpc pc, pc[0x8:e] ; jump to subrutine ++ .long 0 ; Address to static chain, ++ ; filled in by avr32_initialize_trampoline() ++ .long 0 ; Address to subrutine, ++ ; filled in by avr32_initialize_trampoline() ++ */ ++void ++avr32_trampoline_template (FILE * file) ++ { ++ fprintf (file, "\tlddpc r0, pc[8]\n"); ++ fprintf (file, "\tlddpc pc, pc[8]\n"); ++ /* make room for the address of the static chain. */ ++ fprintf (file, "\t.long\t0\n"); ++ /* make room for the address to the subrutine. */ ++ fprintf (file, "\t.long\t0\n"); ++ } ++ ++ ++/* ++ Initialize the variable parts of a trampoline. ++ */ ++void ++avr32_initialize_trampoline (rtx addr, rtx fnaddr, rtx static_chain) ++ { ++ /* Store the address to the static chain. */ ++ emit_move_insn (gen_rtx_MEM ++ (SImode, plus_constant (addr, TRAMPOLINE_SIZE - 4)), ++ static_chain); ++ ++ /* Store the address to the function. */ ++ emit_move_insn (gen_rtx_MEM (SImode, plus_constant (addr, TRAMPOLINE_SIZE)), ++ fnaddr); ++ ++ emit_insn (gen_cache (gen_rtx_REG (SImode, 13), ++ gen_rtx_CONST_INT (SImode, ++ AVR32_CACHE_INVALIDATE_ICACHE))); ++ } ++ ++/* Return nonzero if X is valid as an addressing register. */ ++int ++avr32_address_register_rtx_p (rtx x, int strict_p) ++ { ++ int regno; ++ ++ if (!register_operand(x, GET_MODE(x))) ++ return 0; ++ ++ /* If strict we require the register to be a hard register. */ ++ if (strict_p ++ && !REG_P(x)) ++ return 0; ++ ++ regno = REGNO (x); ++ ++ if (strict_p) ++ return REGNO_OK_FOR_BASE_P (regno); ++ ++ return (regno <= LAST_REGNUM || regno >= FIRST_PSEUDO_REGISTER); ++ } ++ ++/* Return nonzero if INDEX is valid for an address index operand. */ ++int ++avr32_legitimate_index_p (enum machine_mode mode, rtx index, int strict_p) ++ { ++ enum rtx_code code = GET_CODE (index); ++ ++ if (GET_MODE_SIZE (mode) > 8) ++ return 0; ++ ++ /* Standard coprocessor addressing modes. */ ++ if (code == CONST_INT) ++ { ++ if (TARGET_HARD_FLOAT && GET_MODE_CLASS (mode) == MODE_FLOAT) ++ /* Coprocessor mem insns has a smaller reach than ordinary mem insns */ ++ return CONST_OK_FOR_CONSTRAINT_P (INTVAL (index), 'K', "Ku14"); ++ else ++ return CONST_OK_FOR_CONSTRAINT_P (INTVAL (index), 'K', "Ks16"); ++ } ++ ++ if (avr32_address_register_rtx_p (index, strict_p)) ++ return 1; ++ ++ if (code == MULT) ++ { ++ rtx xiop0 = XEXP (index, 0); ++ rtx xiop1 = XEXP (index, 1); ++ return ((avr32_address_register_rtx_p (xiop0, strict_p) ++ && power_of_two_operand (xiop1, SImode) ++ && (INTVAL (xiop1) <= 8)) ++ || (avr32_address_register_rtx_p (xiop1, strict_p) ++ && power_of_two_operand (xiop0, SImode) ++ && (INTVAL (xiop0) <= 8))); ++ } ++ else if (code == ASHIFT) ++ { ++ rtx op = XEXP (index, 1); ++ ++ return (avr32_address_register_rtx_p (XEXP (index, 0), strict_p) ++ && GET_CODE (op) == CONST_INT ++ && INTVAL (op) > 0 && INTVAL (op) <= 3); ++ } ++ ++ return 0; ++ } ++ ++/* ++ Used in the GO_IF_LEGITIMATE_ADDRESS macro. Returns a nonzero value if ++ the RTX x is a legitimate memory address. ++ ++ Returns NO_REGS if the address is not legatime, GENERAL_REGS or ALL_REGS ++ if it is. ++ */ ++ ++/* Forward declaration*/ ++int is_minipool_label (rtx label); ++ ++int ++avr32_legitimate_address (enum machine_mode mode, rtx x, int strict) ++ { ++ ++ switch (GET_CODE (x)) ++ { ++ case REG: ++ return avr32_address_register_rtx_p (x, strict); ++ case CONST: ++ { ++ rtx label = avr32_find_symbol (x); ++ if (label ++ && ++ ((CONSTANT_POOL_ADDRESS_P (label) ++ && !(flag_pic ++ && (symbol_mentioned_p (get_pool_constant (label)) ++ || label_mentioned_p (get_pool_constant (label))))) ++ /* TODO! Can this ever happen??? */ ++ || ((GET_CODE (label) == LABEL_REF) ++ && GET_CODE (XEXP (label, 0)) == CODE_LABEL ++ && is_minipool_label (XEXP (label, 0))))) ++ { ++ return TRUE; ++ } ++ } ++ break; ++ case LABEL_REF: ++ if (GET_CODE (XEXP (x, 0)) == CODE_LABEL ++ && is_minipool_label (XEXP (x, 0))) ++ { ++ return TRUE; ++ } ++ break; ++ case SYMBOL_REF: ++ { ++ if (CONSTANT_POOL_ADDRESS_P (x) ++ && !(flag_pic ++ && (symbol_mentioned_p (get_pool_constant (x)) ++ || label_mentioned_p (get_pool_constant (x))))) ++ return TRUE; ++ /* ++ A symbol_ref is only legal if it is a function. If all of them are ++ legal, a pseudo reg that is a constant will be replaced by a ++ symbol_ref and make illegale code. SYMBOL_REF_FLAG is set by ++ ENCODE_SECTION_INFO. */ ++ else if (SYMBOL_REF_RCALL_FUNCTION_P (x)) ++ return TRUE; ++ break; ++ } ++ case PRE_DEC: /* (pre_dec (...)) */ ++ case POST_INC: /* (post_inc (...)) */ ++ return avr32_address_register_rtx_p (XEXP (x, 0), strict); ++ case PLUS: /* (plus (...) (...)) */ ++ { ++ rtx xop0 = XEXP (x, 0); ++ rtx xop1 = XEXP (x, 1); ++ ++ return ((avr32_address_register_rtx_p (xop0, strict) ++ && avr32_legitimate_index_p (mode, xop1, strict)) ++ || (avr32_address_register_rtx_p (xop1, strict) ++ && avr32_legitimate_index_p (mode, xop0, strict))); ++ } ++ default: ++ break; ++ } ++ ++ return FALSE; ++ } ++ ++ ++int ++avr32_const_double_immediate (rtx value) ++ { ++ HOST_WIDE_INT hi, lo; ++ ++ if (GET_CODE (value) != CONST_DOUBLE) ++ return FALSE; ++ ++ if (SCALAR_FLOAT_MODE_P (GET_MODE (value))) ++ { ++ HOST_WIDE_INT target_float[2]; ++ hi = lo = 0; ++ real_to_target (target_float, CONST_DOUBLE_REAL_VALUE (value), ++ GET_MODE (value)); ++ lo = target_float[0]; ++ hi = target_float[1]; ++ } ++ else ++ { ++ hi = CONST_DOUBLE_HIGH (value); ++ lo = CONST_DOUBLE_LOW (value); ++ } ++ ++ if (avr32_const_ok_for_constraint_p (lo, 'K', "Ks21") ++ && (GET_MODE (value) == SFmode ++ || avr32_const_ok_for_constraint_p (hi, 'K', "Ks21"))) ++ { ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++ ++int ++avr32_legitimate_constant_p (rtx x) ++ { ++ switch (GET_CODE (x)) ++ { ++ case CONST_INT: ++ /* Check if we should put large immediate into constant pool ++ or load them directly with mov/orh.*/ ++ if (!avr32_imm_in_const_pool) ++ return 1; ++ ++ return avr32_const_ok_for_constraint_p (INTVAL (x), 'K', "Ks21"); ++ case CONST_DOUBLE: ++ /* Check if we should put large immediate into constant pool ++ or load them directly with mov/orh.*/ ++ if (!avr32_imm_in_const_pool) ++ return 1; ++ ++ if (GET_MODE (x) == SFmode ++ || GET_MODE (x) == DFmode || GET_MODE (x) == DImode) ++ return avr32_const_double_immediate (x); ++ else ++ return 0; ++ case LABEL_REF: ++ return flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS; ++ case SYMBOL_REF: ++ return flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS; ++ case CONST: ++ case HIGH: ++ case CONST_VECTOR: ++ return 0; ++ default: ++ printf ("%s():\n", __FUNCTION__); ++ debug_rtx (x); ++ return 1; ++ } ++ } ++ ++ ++/* Strip any special encoding from labels */ ++const char * ++avr32_strip_name_encoding (const char *name) ++{ ++ const char *stripped = name; ++ ++ while (1) ++ { ++ switch (stripped[0]) ++ { ++ case '#': ++ stripped = strchr (name + 1, '#') + 1; ++ break; ++ case '*': ++ stripped = &stripped[1]; ++ break; ++ default: ++ return stripped; ++ } ++ } ++} ++ ++ ++ ++/* Do anything needed before RTL is emitted for each function. */ ++static struct machine_function * ++avr32_init_machine_status (void) ++{ ++ struct machine_function *machine; ++ machine = ++ (machine_function *) ggc_alloc_cleared (sizeof (machine_function)); ++ ++#if AVR32_FT_UNKNOWN != 0 ++ machine->func_type = AVR32_FT_UNKNOWN; ++#endif ++ ++ machine->minipool_label_head = 0; ++ machine->minipool_label_tail = 0; ++ return machine; ++} ++ ++void ++avr32_init_expanders (void) ++ { ++ /* Arrange to initialize and mark the machine per-function status. */ ++ init_machine_status = avr32_init_machine_status; ++ } ++ ++ ++/* Return an RTX indicating where the return address to the ++ calling function can be found. */ ++ ++rtx ++avr32_return_addr (int count, rtx frame ATTRIBUTE_UNUSED) ++ { ++ if (count != 0) ++ return NULL_RTX; ++ ++ return get_hard_reg_initial_val (Pmode, LR_REGNUM); ++ } ++ ++ ++void ++avr32_encode_section_info (tree decl, rtx rtl, int first) ++ { ++ ++ if (first && DECL_P (decl)) ++ { ++ /* Set SYMBOL_REG_FLAG for local functions */ ++ if (!TREE_PUBLIC (decl) && TREE_CODE (decl) == FUNCTION_DECL) ++ { ++ if ((*targetm.binds_local_p) (decl)) ++ { ++ SYMBOL_REF_FLAG (XEXP (rtl, 0)) = 1; ++ } ++ } ++ } ++ } ++ ++ ++void ++avr32_asm_output_ascii (FILE * stream, char *ptr, int len) ++ { ++ int i, i_new = 0; ++ char *new_ptr = xmalloc (4 * len); ++ if (new_ptr == NULL) ++ internal_error ("Out of memory."); ++ ++ for (i = 0; i < len; i++) ++ { ++ if (ptr[i] == '\n') ++ { ++ new_ptr[i_new++] = '\\'; ++ new_ptr[i_new++] = '0'; ++ new_ptr[i_new++] = '1'; ++ new_ptr[i_new++] = '2'; ++ } ++ else if (ptr[i] == '\"') ++ { ++ new_ptr[i_new++] = '\\'; ++ new_ptr[i_new++] = '\"'; ++ } ++ else if (ptr[i] == '\\') ++ { ++ new_ptr[i_new++] = '\\'; ++ new_ptr[i_new++] = '\\'; ++ } ++ else if (ptr[i] == '\0' && i + 1 < len) ++ { ++ new_ptr[i_new++] = '\\'; ++ new_ptr[i_new++] = '0'; ++ } ++ else ++ { ++ new_ptr[i_new++] = ptr[i]; ++ } ++ } ++ ++ /* Terminate new_ptr. */ ++ new_ptr[i_new] = '\0'; ++ fprintf (stream, "\t.ascii\t\"%s\"\n", new_ptr); ++ free (new_ptr); ++ } ++ ++ ++void ++avr32_asm_output_label (FILE * stream, const char *name) ++ { ++ name = avr32_strip_name_encoding (name); ++ ++ /* Print the label. */ ++ assemble_name (stream, name); ++ fprintf (stream, ":\n"); ++ } ++ ++ ++ ++void ++avr32_asm_weaken_label (FILE * stream, const char *name) ++ { ++ fprintf (stream, "\t.weak "); ++ assemble_name (stream, name); ++ fprintf (stream, "\n"); ++ } ++ ++/* ++ Checks if a labelref is equal to a reserved word in the assembler. If it is, ++ insert a '_' before the label name. ++ */ ++void ++avr32_asm_output_labelref (FILE * stream, const char *name) ++ { ++ int verbatim = FALSE; ++ const char *stripped = name; ++ int strip_finished = FALSE; ++ ++ while (!strip_finished) ++ { ++ switch (stripped[0]) ++ { ++ case '#': ++ stripped = strchr (name + 1, '#') + 1; ++ break; ++ case '*': ++ stripped = &stripped[1]; ++ verbatim = TRUE; ++ break; ++ default: ++ strip_finished = TRUE; ++ break; ++ } ++ } ++ ++ if (verbatim) ++ fputs (stripped, stream); ++ else ++ asm_fprintf (stream, "%U%s", stripped); ++ } ++ ++ ++ ++/* ++ Check if the comparison in compare_exp is redundant ++ for the condition given in next_cond given that the ++ needed flags are already set by an earlier instruction. ++ Uses cc_prev_status to check this. ++ ++ Returns NULL_RTX if the compare is not redundant ++ or the new condition to use in the conditional ++ instruction if the compare is redundant. ++ */ ++static rtx ++is_compare_redundant (rtx compare_exp, rtx next_cond) ++ { ++ int z_flag_valid = FALSE; ++ int n_flag_valid = FALSE; ++ rtx new_cond; ++ ++ if (GET_CODE (compare_exp) != COMPARE) ++ return NULL_RTX; ++ ++ ++ if (rtx_equal_p (cc_prev_status.mdep.value, compare_exp)) ++ { ++ /* cc0 already contains the correct comparison -> delete cmp insn */ ++ return next_cond; ++ } ++ ++ if (GET_MODE (compare_exp) != SImode) ++ return NULL_RTX; ++ ++ switch (cc_prev_status.mdep.flags) ++ { ++ case CC_SET_VNCZ: ++ case CC_SET_NCZ: ++ n_flag_valid = TRUE; ++ case CC_SET_CZ: ++ case CC_SET_Z: ++ z_flag_valid = TRUE; ++ } ++ ++ if (cc_prev_status.mdep.value ++ && REG_P (XEXP (compare_exp, 0)) ++ && REGNO (XEXP (compare_exp, 0)) == REGNO (cc_prev_status.mdep.value) ++ && GET_CODE (XEXP (compare_exp, 1)) == CONST_INT ++ && next_cond != NULL_RTX) ++ { ++ if (INTVAL (XEXP (compare_exp, 1)) == 0 ++ && z_flag_valid ++ && (GET_CODE (next_cond) == EQ || GET_CODE (next_cond) == NE)) ++ /* We can skip comparison Z flag is already reflecting ops[0] */ ++ return next_cond; ++ else if (n_flag_valid ++ && ((INTVAL (XEXP (compare_exp, 1)) == 0 ++ && (GET_CODE (next_cond) == GE ++ || GET_CODE (next_cond) == LT)) ++ || (INTVAL (XEXP (compare_exp, 1)) == -1 ++ && (GET_CODE (next_cond) == GT ++ || GET_CODE (next_cond) == LE)))) ++ { ++ /* We can skip comparison N flag is already reflecting ops[0], ++ which means that we can use the mi/pl conditions to check if ++ ops[0] is GE or LT 0. */ ++ if ((GET_CODE (next_cond) == GE) || (GET_CODE (next_cond) == GT)) ++ new_cond = ++ gen_rtx_UNSPEC (GET_MODE (next_cond), gen_rtvec (2, cc0_rtx, const0_rtx), ++ UNSPEC_COND_PL); ++ else ++ new_cond = ++ gen_rtx_UNSPEC (GET_MODE (next_cond), gen_rtvec (2, cc0_rtx, const0_rtx), ++ UNSPEC_COND_MI); ++ return new_cond; ++ } ++ } ++ return NULL_RTX; ++ } ++ ++/* Updates cc_status. */ ++void ++avr32_notice_update_cc (rtx exp, rtx insn) ++ { ++ switch (get_attr_cc (insn)) ++ { ++ case CC_CALL_SET: ++ CC_STATUS_INIT; ++ FPCC_STATUS_INIT; ++ /* Check if the function call returns a value in r12 */ ++ if (REG_P (recog_data.operand[0]) ++ && REGNO (recog_data.operand[0]) == RETVAL_REGNUM) ++ { ++ cc_status.flags = 0; ++ cc_status.mdep.value = ++ gen_rtx_COMPARE (SImode, recog_data.operand[0], const0_rtx); ++ cc_status.mdep.flags = CC_SET_VNCZ; ++ ++ } ++ break; ++ case CC_COMPARE: ++ /* Check that compare will not be optimized away if so nothing should ++ be done */ ++ if (is_compare_redundant (SET_SRC (exp), get_next_insn_cond (insn)) == ++ NULL_RTX) ++ { ++ ++ /* Reset the nonstandard flag */ ++ CC_STATUS_INIT; ++ cc_status.flags = 0; ++ cc_status.mdep.value = SET_SRC (exp); ++ cc_status.mdep.flags = CC_SET_VNCZ; ++ } ++ break; ++ case CC_CMP_COND_INSN: ++ { ++ /* Conditional insn that emit the compare itself. */ ++ rtx cmp = gen_rtx_COMPARE (GET_MODE (recog_data.operand[4]), ++ recog_data.operand[4], ++ recog_data.operand[5]); ++ ++ if (is_compare_redundant (cmp, recog_data.operand[1]) == NULL_RTX) ++ { ++ ++ /* Reset the nonstandard flag */ ++ CC_STATUS_INIT; ++ cc_status.flags = 0; ++ cc_status.mdep.value = cmp; ++ cc_status.mdep.flags = CC_SET_VNCZ; ++ } ++ } ++ break; ++ case CC_FPCOMPARE: ++ /* Check that floating-point compare will not be optimized away if so ++ nothing should be done */ ++ if (!rtx_equal_p (cc_prev_status.mdep.fpvalue, SET_SRC (exp))) ++ { ++ /* cc0 already contains the correct comparison -> delete cmp insn */ ++ /* Reset the nonstandard flag */ ++ cc_status.mdep.fpvalue = SET_SRC (exp); ++ cc_status.mdep.fpflags = CC_SET_CZ; ++ } ++ break; ++ case CC_FROM_FPCC: ++ /* Flags are updated with flags from Floating-point coprocessor, set ++ CC_NOT_SIGNED flag since the flags are set so that unsigned ++ condidion codes can be used directly. */ ++ CC_STATUS_INIT; ++ cc_status.flags = CC_NOT_SIGNED; ++ cc_status.mdep.value = cc_status.mdep.fpvalue; ++ cc_status.mdep.flags = cc_status.mdep.fpflags; ++ break; ++ case CC_BLD: ++ /* Bit load is kind of like an inverted testsi, because the Z flag is ++ inverted */ ++ CC_STATUS_INIT; ++ cc_status.flags = CC_INVERTED; ++ cc_status.mdep.value = SET_SRC (exp); ++ cc_status.mdep.flags = CC_SET_Z; ++ break; ++ case CC_NONE: ++ /* Insn does not affect CC at all. Check if the instruction updates ++ some of the register currently reflected in cc0 */ ++ ++ if ((GET_CODE (exp) == SET) ++ && (cc_status.value1 || cc_status.value2 || cc_status.mdep.value) ++ && (reg_mentioned_p (SET_DEST (exp), cc_status.value1) ++ || reg_mentioned_p (SET_DEST (exp), cc_status.value2) ++ || reg_mentioned_p (SET_DEST (exp), cc_status.mdep.value))) ++ { ++ CC_STATUS_INIT; ++ } ++ ++ /* If this is a parallel we must step through each of the parallel ++ expressions */ ++ if (GET_CODE (exp) == PARALLEL) ++ { ++ int i; ++ for (i = 0; i < XVECLEN (exp, 0); ++i) ++ { ++ rtx vec_exp = XVECEXP (exp, 0, i); ++ if ((GET_CODE (vec_exp) == SET) ++ && (cc_status.value1 || cc_status.value2 ++ || cc_status.mdep.value) ++ && (reg_mentioned_p (SET_DEST (vec_exp), cc_status.value1) ++ || reg_mentioned_p (SET_DEST (vec_exp), ++ cc_status.value2) ++ || reg_mentioned_p (SET_DEST (vec_exp), ++ cc_status.mdep.value))) ++ { ++ CC_STATUS_INIT; ++ } ++ } ++ } ++ ++ /* Check if we have memory opartions with post_inc or pre_dec on the ++ register currently reflected in cc0 */ ++ if (GET_CODE (exp) == SET ++ && GET_CODE (SET_SRC (exp)) == MEM ++ && (GET_CODE (XEXP (SET_SRC (exp), 0)) == POST_INC ++ || GET_CODE (XEXP (SET_SRC (exp), 0)) == PRE_DEC) ++ && ++ (reg_mentioned_p ++ (XEXP (XEXP (SET_SRC (exp), 0), 0), cc_status.value1) ++ || reg_mentioned_p (XEXP (XEXP (SET_SRC (exp), 0), 0), ++ cc_status.value2) ++ || reg_mentioned_p (XEXP (XEXP (SET_SRC (exp), 0), 0), ++ cc_status.mdep.value))) ++ CC_STATUS_INIT; ++ ++ if (GET_CODE (exp) == SET ++ && GET_CODE (SET_DEST (exp)) == MEM ++ && (GET_CODE (XEXP (SET_DEST (exp), 0)) == POST_INC ++ || GET_CODE (XEXP (SET_DEST (exp), 0)) == PRE_DEC) ++ && ++ (reg_mentioned_p ++ (XEXP (XEXP (SET_DEST (exp), 0), 0), cc_status.value1) ++ || reg_mentioned_p (XEXP (XEXP (SET_DEST (exp), 0), 0), ++ cc_status.value2) ++ || reg_mentioned_p (XEXP (XEXP (SET_DEST (exp), 0), 0), ++ cc_status.mdep.value))) ++ CC_STATUS_INIT; ++ break; ++ ++ case CC_SET_VNCZ: ++ CC_STATUS_INIT; ++ cc_status.mdep.value = recog_data.operand[0]; ++ cc_status.mdep.flags = CC_SET_VNCZ; ++ break; ++ ++ case CC_SET_NCZ: ++ CC_STATUS_INIT; ++ cc_status.mdep.value = recog_data.operand[0]; ++ cc_status.mdep.flags = CC_SET_NCZ; ++ break; ++ ++ case CC_SET_CZ: ++ CC_STATUS_INIT; ++ cc_status.mdep.value = recog_data.operand[0]; ++ cc_status.mdep.flags = CC_SET_CZ; ++ break; ++ ++ case CC_SET_Z: ++ CC_STATUS_INIT; ++ cc_status.mdep.value = recog_data.operand[0]; ++ cc_status.mdep.flags = CC_SET_Z; ++ break; ++ ++ case CC_CLOBBER: ++ CC_STATUS_INIT; ++ break; ++ ++ default: ++ CC_STATUS_INIT; ++ } ++ } ++ ++ ++/* ++ Outputs to stdio stream stream the assembler syntax for an instruction ++ operand x. x is an RTL expression. ++ */ ++void ++avr32_print_operand (FILE * stream, rtx x, int code) ++ { ++ int error = 0; ++ ++ switch (GET_CODE (x)) ++ { ++ case UNSPEC: ++ switch (XINT (x, 1)) ++ { ++ case UNSPEC_COND_PL: ++ if (code == 'i') ++ fputs ("mi", stream); ++ else ++ fputs ("pl", stream); ++ break; ++ case UNSPEC_COND_MI: ++ if (code == 'i') ++ fputs ("pl", stream); ++ else ++ fputs ("mi", stream); ++ break; ++ default: ++ error = 1; ++ } ++ break; ++ case EQ: ++ if (code == 'i') ++ fputs ("ne", stream); ++ else ++ fputs ("eq", stream); ++ break; ++ case NE: ++ if (code == 'i') ++ fputs ("eq", stream); ++ else ++ fputs ("ne", stream); ++ break; ++ case GT: ++ if (code == 'i') ++ fputs ("le", stream); ++ else ++ fputs ("gt", stream); ++ break; ++ case GTU: ++ if (code == 'i') ++ fputs ("ls", stream); ++ else ++ fputs ("hi", stream); ++ break; ++ case LT: ++ if (code == 'i') ++ fputs ("ge", stream); ++ else ++ fputs ("lt", stream); ++ break; ++ case LTU: ++ if (code == 'i') ++ fputs ("hs", stream); ++ else ++ fputs ("lo", stream); ++ break; ++ case GE: ++ if (code == 'i') ++ fputs ("lt", stream); ++ else ++ fputs ("ge", stream); ++ break; ++ case GEU: ++ if (code == 'i') ++ fputs ("lo", stream); ++ else ++ fputs ("hs", stream); ++ break; ++ case LE: ++ if (code == 'i') ++ fputs ("gt", stream); ++ else ++ fputs ("le", stream); ++ break; ++ case LEU: ++ if (code == 'i') ++ fputs ("hi", stream); ++ else ++ fputs ("ls", stream); ++ break; ++ case CONST_INT: ++ { ++ HOST_WIDE_INT value = INTVAL (x); ++ ++ switch (code) ++ { ++ case 'm': ++ if ( HOST_BITS_PER_WIDE_INT > BITS_PER_WORD ) ++ { ++ /* A const_int can be used to represent DImode constants. */ ++ value >>= BITS_PER_WORD; ++ } ++ /* We might get a const_int immediate for setting a DI register, ++ we then must then return the correct sign extended DI. The most ++ significant word is just a sign extension. */ ++ else if (value < 0) ++ value = -1; ++ else ++ value = 0; ++ break; ++ case 'i': ++ value++; ++ break; ++ case 'p': ++ { ++ /* Set to bit position of first bit set in immediate */ ++ int i, bitpos = 32; ++ for (i = 0; i < 32; i++) ++ if (value & (1 << i)) ++ { ++ bitpos = i; ++ break; ++ } ++ value = bitpos; ++ } ++ break; ++ case 'r': ++ { ++ /* Reglist 8 */ ++ char op[50]; ++ op[0] = '\0'; ++ ++ if (value & 0x01) ++ sprintf (op, "r0-r3"); ++ if (value & 0x02) ++ strlen (op) ? sprintf (op, "%s, r4-r7", op) : sprintf (op, ++ "r4-r7"); ++ if (value & 0x04) ++ strlen (op) ? sprintf (op, "%s, r8-r9", op) : sprintf (op, ++ "r8-r9"); ++ if (value & 0x08) ++ strlen (op) ? sprintf (op, "%s, r10", op) : sprintf (op, ++ "r10"); ++ if (value & 0x10) ++ strlen (op) ? sprintf (op, "%s, r11", op) : sprintf (op, ++ "r11"); ++ if (value & 0x20) ++ strlen (op) ? sprintf (op, "%s, r12", op) : sprintf (op, ++ "r12"); ++ if (value & 0x40) ++ strlen (op) ? sprintf (op, "%s, lr", op) : sprintf (op, "lr"); ++ if (value & 0x80) ++ strlen (op) ? sprintf (op, "%s, pc", op) : sprintf (op, "pc"); ++ ++ fputs (op, stream); ++ return; ++ } ++ case 's': ++ { ++ /* Reglist 16 */ ++ char reglist16_string[100]; ++ int i; ++ reglist16_string[0] = '\0'; ++ ++ for (i = 0; i < 16; ++i) ++ { ++ if (value & (1 << i)) ++ { ++ strlen (reglist16_string) ? sprintf (reglist16_string, ++ "%s, %s", ++ reglist16_string, ++ reg_names ++ [INTERNAL_REGNUM ++ (i)]) : ++ sprintf (reglist16_string, "%s", ++ reg_names[INTERNAL_REGNUM (i)]); ++ } ++ } ++ fputs (reglist16_string, stream); ++ return; ++ } ++ case 'C': ++ { ++ /* RegListCP8 */ ++ char reglist_string[100]; ++ avr32_make_fp_reglist_w (value, (char *) reglist_string); ++ fputs (reglist_string, stream); ++ return; ++ } ++ case 'D': ++ { ++ /* RegListCPD8 */ ++ char reglist_string[100]; ++ avr32_make_fp_reglist_d (value, (char *) reglist_string); ++ fputs (reglist_string, stream); ++ return; ++ } ++ case 'h': ++ /* Print halfword part of word */ ++ fputs (value ? "b" : "t", stream); ++ return; ++ } ++ ++ /* Print Value */ ++ fprintf (stream, "%d", value); ++ break; ++ } ++ case CONST_DOUBLE: ++ { ++ HOST_WIDE_INT hi, lo; ++ if (SCALAR_FLOAT_MODE_P (GET_MODE (x))) ++ { ++ HOST_WIDE_INT target_float[2]; ++ hi = lo = 0; ++ real_to_target (target_float, CONST_DOUBLE_REAL_VALUE (x), ++ GET_MODE (x)); ++ /* For doubles the most significant part starts at index 0. */ ++ if (GET_MODE_SIZE (GET_MODE (x)) > UNITS_PER_WORD) ++ { ++ hi = target_float[0]; ++ lo = target_float[1]; ++ } ++ else ++ { ++ lo = target_float[0]; ++ } ++ } ++ else ++ { ++ hi = CONST_DOUBLE_HIGH (x); ++ lo = CONST_DOUBLE_LOW (x); ++ } ++ ++ if (code == 'm') ++ fprintf (stream, "%ld", hi); ++ else ++ fprintf (stream, "%ld", lo); ++ ++ break; ++ } ++ case CONST: ++ output_addr_const (stream, XEXP (XEXP (x, 0), 0)); ++ fprintf (stream, "+%ld", INTVAL (XEXP (XEXP (x, 0), 1))); ++ break; ++ case REG: ++ /* Swap register name if the register is DImode or DFmode. */ ++ if (GET_MODE (x) == DImode || GET_MODE (x) == DFmode) ++ { ++ /* Double register must have an even numbered address */ ++ gcc_assert (!(REGNO (x) % 2)); ++ if (code == 'm') ++ fputs (reg_names[true_regnum (x)], stream); ++ else ++ fputs (reg_names[true_regnum (x) + 1], stream); ++ } ++ else if (GET_MODE (x) == TImode) ++ { ++ switch (code) ++ { ++ case 'T': ++ fputs (reg_names[true_regnum (x)], stream); ++ break; ++ case 'U': ++ fputs (reg_names[true_regnum (x) + 1], stream); ++ break; ++ case 'L': ++ fputs (reg_names[true_regnum (x) + 2], stream); ++ break; ++ case 'B': ++ fputs (reg_names[true_regnum (x) + 3], stream); ++ break; ++ default: ++ fprintf (stream, "%s, %s, %s, %s", ++ reg_names[true_regnum (x) + 3], ++ reg_names[true_regnum (x) + 2], ++ reg_names[true_regnum (x) + 1], ++ reg_names[true_regnum (x)]); ++ break; ++ } ++ } ++ else ++ { ++ fputs (reg_names[true_regnum (x)], stream); ++ } ++ break; ++ case CODE_LABEL: ++ case LABEL_REF: ++ case SYMBOL_REF: ++ output_addr_const (stream, x); ++ break; ++ case MEM: ++ switch (GET_CODE (XEXP (x, 0))) ++ { ++ case LABEL_REF: ++ case SYMBOL_REF: ++ output_addr_const (stream, XEXP (x, 0)); ++ break; ++ case MEM: ++ switch (GET_CODE (XEXP (XEXP (x, 0), 0))) ++ { ++ case SYMBOL_REF: ++ output_addr_const (stream, XEXP (XEXP (x, 0), 0)); ++ break; ++ default: ++ error = 1; ++ break; ++ } ++ break; ++ case REG: ++ avr32_print_operand (stream, XEXP (x, 0), 0); ++ if (code != 'p') ++ fputs ("[0]", stream); ++ break; ++ case PRE_DEC: ++ fputs ("--", stream); ++ avr32_print_operand (stream, XEXP (XEXP (x, 0), 0), 0); ++ break; ++ case POST_INC: ++ avr32_print_operand (stream, XEXP (XEXP (x, 0), 0), 0); ++ fputs ("++", stream); ++ break; ++ case PLUS: ++ { ++ rtx op0 = XEXP (XEXP (x, 0), 0); ++ rtx op1 = XEXP (XEXP (x, 0), 1); ++ rtx base = NULL_RTX, offset = NULL_RTX; ++ ++ if (avr32_address_register_rtx_p (op0, 1)) ++ { ++ base = op0; ++ offset = op1; ++ } ++ else if (avr32_address_register_rtx_p (op1, 1)) ++ { ++ /* Operands are switched. */ ++ base = op1; ++ offset = op0; ++ } ++ ++ gcc_assert (base && offset ++ && avr32_address_register_rtx_p (base, 1) ++ && avr32_legitimate_index_p (GET_MODE (x), offset, ++ 1)); ++ ++ avr32_print_operand (stream, base, 0); ++ fputs ("[", stream); ++ avr32_print_operand (stream, offset, 0); ++ fputs ("]", stream); ++ break; ++ } ++ case CONST: ++ output_addr_const (stream, XEXP (XEXP (XEXP (x, 0), 0), 0)); ++ fprintf (stream, " + %ld", ++ INTVAL (XEXP (XEXP (XEXP (x, 0), 0), 1))); ++ break; ++ default: ++ error = 1; ++ } ++ break; ++ case MULT: ++ { ++ int value = INTVAL (XEXP (x, 1)); ++ ++ /* Convert immediate in multiplication into a shift immediate */ ++ switch (value) ++ { ++ case 2: ++ value = 1; ++ break; ++ case 4: ++ value = 2; ++ break; ++ case 8: ++ value = 3; ++ break; ++ default: ++ value = 0; ++ } ++ fprintf (stream, "%s << %i", reg_names[true_regnum (XEXP (x, 0))], ++ value); ++ break; ++ } ++ case ASHIFT: ++ if (GET_CODE (XEXP (x, 1)) == CONST_INT) ++ fprintf (stream, "%s << %i", reg_names[true_regnum (XEXP (x, 0))], ++ (int) INTVAL (XEXP (x, 1))); ++ else if (REG_P (XEXP (x, 1))) ++ fprintf (stream, "%s << %s", reg_names[true_regnum (XEXP (x, 0))], ++ reg_names[true_regnum (XEXP (x, 1))]); ++ else ++ { ++ error = 1; ++ } ++ break; ++ case LSHIFTRT: ++ if (GET_CODE (XEXP (x, 1)) == CONST_INT) ++ fprintf (stream, "%s >> %i", reg_names[true_regnum (XEXP (x, 0))], ++ (int) INTVAL (XEXP (x, 1))); ++ else if (REG_P (XEXP (x, 1))) ++ fprintf (stream, "%s >> %s", reg_names[true_regnum (XEXP (x, 0))], ++ reg_names[true_regnum (XEXP (x, 1))]); ++ else ++ { ++ error = 1; ++ } ++ fprintf (stream, ">>"); ++ break; ++ case PARALLEL: ++ { ++ /* Load store multiple */ ++ int i; ++ int count = XVECLEN (x, 0); ++ int reglist16 = 0; ++ char reglist16_string[100]; ++ ++ for (i = 0; i < count; ++i) ++ { ++ rtx vec_elm = XVECEXP (x, 0, i); ++ if (GET_MODE (vec_elm) != SET) ++ { ++ debug_rtx (vec_elm); ++ internal_error ("Unknown element in parallel expression!"); ++ } ++ if (GET_MODE (XEXP (vec_elm, 0)) == REG) ++ { ++ /* Load multiple */ ++ reglist16 |= 1 << ASM_REGNUM (REGNO (XEXP (vec_elm, 0))); ++ } ++ else ++ { ++ /* Store multiple */ ++ reglist16 |= 1 << ASM_REGNUM (REGNO (XEXP (vec_elm, 1))); ++ } ++ } ++ ++ avr32_make_reglist16 (reglist16, reglist16_string); ++ fputs (reglist16_string, stream); ++ ++ break; ++ } ++ ++ default: ++ error = 1; ++ } ++ ++ if (error) ++ { ++ debug_rtx (x); ++ internal_error ("Illegal expression for avr32_print_operand"); ++ } ++ } ++ ++rtx ++avr32_get_note_reg_equiv (rtx insn) ++ { ++ rtx note; ++ ++ note = find_reg_note (insn, REG_EQUIV, NULL_RTX); ++ ++ if (note != NULL_RTX) ++ return XEXP (note, 0); ++ else ++ return NULL_RTX; ++ } ++ ++/* ++ Outputs to stdio stream stream the assembler syntax for an instruction ++ operand that is a memory reference whose address is x. x is an RTL ++ expression. ++ ++ ToDo: fixme. ++ */ ++void ++avr32_print_operand_address (FILE * stream, rtx x) ++ { ++ fprintf (stream, "(%d) /* address */", REGNO (x)); ++ } ++ ++/* Return true if _GLOBAL_OFFSET_TABLE_ symbol is mentioned. */ ++bool ++avr32_got_mentioned_p (rtx addr) ++ { ++ if (GET_CODE (addr) == MEM) ++ addr = XEXP (addr, 0); ++ while (GET_CODE (addr) == CONST) ++ addr = XEXP (addr, 0); ++ if (GET_CODE (addr) == SYMBOL_REF) ++ { ++ return streq (XSTR (addr, 0), "_GLOBAL_OFFSET_TABLE_"); ++ } ++ if (GET_CODE (addr) == PLUS || GET_CODE (addr) == MINUS) ++ { ++ bool l1, l2; ++ ++ l1 = avr32_got_mentioned_p (XEXP (addr, 0)); ++ l2 = avr32_got_mentioned_p (XEXP (addr, 1)); ++ return l1 || l2; ++ } ++ return false; ++ } ++ ++ ++/* Find the symbol in an address expression. */ ++ ++rtx ++avr32_find_symbol (rtx addr) ++ { ++ if (GET_CODE (addr) == MEM) ++ addr = XEXP (addr, 0); ++ ++ while (GET_CODE (addr) == CONST) ++ addr = XEXP (addr, 0); ++ ++ if (GET_CODE (addr) == SYMBOL_REF || GET_CODE (addr) == LABEL_REF) ++ return addr; ++ if (GET_CODE (addr) == PLUS) ++ { ++ rtx l1, l2; ++ ++ l1 = avr32_find_symbol (XEXP (addr, 0)); ++ l2 = avr32_find_symbol (XEXP (addr, 1)); ++ if (l1 != NULL_RTX && l2 == NULL_RTX) ++ return l1; ++ else if (l1 == NULL_RTX && l2 != NULL_RTX) ++ return l2; ++ } ++ ++ return NULL_RTX; ++ } ++ ++ ++/* Routines for manipulation of the constant pool. */ ++ ++/* AVR32 instructions cannot load a large constant directly into a ++ register; they have to come from a pc relative load. The constant ++ must therefore be placed in the addressable range of the pc ++ relative load. Depending on the precise pc relative load ++ instruction the range is somewhere between 256 bytes and 4k. This ++ means that we often have to dump a constant inside a function, and ++ generate code to branch around it. ++ ++ It is important to minimize this, since the branches will slow ++ things down and make the code larger. ++ ++ Normally we can hide the table after an existing unconditional ++ branch so that there is no interruption of the flow, but in the ++ worst case the code looks like this: ++ ++ lddpc rn, L1 ++ ... ++ rjmp L2 ++ align ++ L1: .long value ++ L2: ++ ... ++ ++ lddpc rn, L3 ++ ... ++ rjmp L4 ++ align ++ L3: .long value ++ L4: ++ ... ++ ++ We fix this by performing a scan after scheduling, which notices ++ which instructions need to have their operands fetched from the ++ constant table and builds the table. ++ ++ The algorithm starts by building a table of all the constants that ++ need fixing up and all the natural barriers in the function (places ++ where a constant table can be dropped without breaking the flow). ++ For each fixup we note how far the pc-relative replacement will be ++ able to reach and the offset of the instruction into the function. ++ ++ Having built the table we then group the fixes together to form ++ tables that are as large as possible (subject to addressing ++ constraints) and emit each table of constants after the last ++ barrier that is within range of all the instructions in the group. ++ If a group does not contain a barrier, then we forcibly create one ++ by inserting a jump instruction into the flow. Once the table has ++ been inserted, the insns are then modified to reference the ++ relevant entry in the pool. ++ ++ Possible enhancements to the algorithm (not implemented) are: ++ ++ 1) For some processors and object formats, there may be benefit in ++ aligning the pools to the start of cache lines; this alignment ++ would need to be taken into account when calculating addressability ++ of a pool. */ ++ ++/* These typedefs are located at the start of this file, so that ++ they can be used in the prototypes there. This comment is to ++ remind readers of that fact so that the following structures ++ can be understood more easily. ++ ++ typedef struct minipool_node Mnode; ++ typedef struct minipool_fixup Mfix; */ ++ ++struct minipool_node ++{ ++ /* Doubly linked chain of entries. */ ++ Mnode *next; ++ Mnode *prev; ++ /* The maximum offset into the code that this entry can be placed. While ++ pushing fixes for forward references, all entries are sorted in order of ++ increasing max_address. */ ++ HOST_WIDE_INT max_address; ++ /* Similarly for an entry inserted for a backwards ref. */ ++ HOST_WIDE_INT min_address; ++ /* The number of fixes referencing this entry. This can become zero if we ++ "unpush" an entry. In this case we ignore the entry when we come to ++ emit the code. */ ++ int refcount; ++ /* The offset from the start of the minipool. */ ++ HOST_WIDE_INT offset; ++ /* The value in table. */ ++ rtx value; ++ /* The mode of value. */ ++ enum machine_mode mode; ++ /* The size of the value. */ ++ int fix_size; ++}; ++ ++struct minipool_fixup ++{ ++ Mfix *next; ++ rtx insn; ++ HOST_WIDE_INT address; ++ rtx *loc; ++ enum machine_mode mode; ++ int fix_size; ++ rtx value; ++ Mnode *minipool; ++ HOST_WIDE_INT forwards; ++ HOST_WIDE_INT backwards; ++}; ++ ++ ++/* Fixes less than a word need padding out to a word boundary. */ ++#define MINIPOOL_FIX_SIZE(mode, value) \ ++ (IS_FORCE_MINIPOOL(value) ? 0 : \ ++ (GET_MODE_SIZE ((mode)) >= 4 ? GET_MODE_SIZE ((mode)) : 4)) ++ ++#define IS_FORCE_MINIPOOL(x) \ ++ (GET_CODE(x) == UNSPEC && \ ++ XINT(x, 1) == UNSPEC_FORCE_MINIPOOL) ++ ++static Mnode *minipool_vector_head; ++static Mnode *minipool_vector_tail; ++ ++/* The linked list of all minipool fixes required for this function. */ ++Mfix *minipool_fix_head; ++Mfix *minipool_fix_tail; ++/* The fix entry for the current minipool, once it has been placed. */ ++Mfix *minipool_barrier; ++ ++/* Determines if INSN is the start of a jump table. Returns the end ++ of the TABLE or NULL_RTX. */ ++static rtx ++is_jump_table (rtx insn) ++ { ++ rtx table; ++ ++ if (GET_CODE (insn) == JUMP_INSN ++ && JUMP_LABEL (insn) != NULL ++ && ((table = next_real_insn (JUMP_LABEL (insn))) ++ == next_real_insn (insn)) ++ && table != NULL ++ && GET_CODE (table) == JUMP_INSN ++ && (GET_CODE (PATTERN (table)) == ADDR_VEC ++ || GET_CODE (PATTERN (table)) == ADDR_DIFF_VEC)) ++ return table; ++ ++ return NULL_RTX; ++ } ++ ++static HOST_WIDE_INT ++get_jump_table_size (rtx insn) ++ { ++ /* ADDR_VECs only take room if read-only data does into the text section. */ ++ if (JUMP_TABLES_IN_TEXT_SECTION ++#if !defined(READONLY_DATA_SECTION_ASM_OP) ++ || 1 ++#endif ++ ) ++ { ++ rtx body = PATTERN (insn); ++ int elt = GET_CODE (body) == ADDR_DIFF_VEC ? 1 : 0; ++ ++ return GET_MODE_SIZE (GET_MODE (body)) * XVECLEN (body, elt); ++ } ++ ++ return 0; ++ } ++ ++/* Move a minipool fix MP from its current location to before MAX_MP. ++ If MAX_MP is NULL, then MP doesn't need moving, but the addressing ++ constraints may need updating. */ ++static Mnode * ++move_minipool_fix_forward_ref (Mnode * mp, Mnode * max_mp, ++ HOST_WIDE_INT max_address) ++ { ++ /* This should never be true and the code below assumes these are ++ different. */ ++ if (mp == max_mp) ++ abort (); ++ ++ if (max_mp == NULL) ++ { ++ if (max_address < mp->max_address) ++ mp->max_address = max_address; ++ } ++ else ++ { ++ if (max_address > max_mp->max_address - mp->fix_size) ++ mp->max_address = max_mp->max_address - mp->fix_size; ++ else ++ mp->max_address = max_address; ++ ++ /* Unlink MP from its current position. Since max_mp is non-null, ++ mp->prev must be non-null. */ ++ mp->prev->next = mp->next; ++ if (mp->next != NULL) ++ mp->next->prev = mp->prev; ++ else ++ minipool_vector_tail = mp->prev; ++ ++ /* Re-insert it before MAX_MP. */ ++ mp->next = max_mp; ++ mp->prev = max_mp->prev; ++ max_mp->prev = mp; ++ ++ if (mp->prev != NULL) ++ mp->prev->next = mp; ++ else ++ minipool_vector_head = mp; ++ } ++ ++ /* Save the new entry. */ ++ max_mp = mp; ++ ++ /* Scan over the preceding entries and adjust their addresses as required. ++ */ ++ while (mp->prev != NULL ++ && mp->prev->max_address > mp->max_address - mp->prev->fix_size) ++ { ++ mp->prev->max_address = mp->max_address - mp->prev->fix_size; ++ mp = mp->prev; ++ } ++ ++ return max_mp; ++ } ++ ++/* Add a constant to the minipool for a forward reference. Returns the ++ node added or NULL if the constant will not fit in this pool. */ ++static Mnode * ++add_minipool_forward_ref (Mfix * fix) ++{ ++ /* If set, max_mp is the first pool_entry that has a lower constraint than ++ the one we are trying to add. */ ++ Mnode *max_mp = NULL; ++ HOST_WIDE_INT max_address = fix->address + fix->forwards; ++ Mnode *mp; ++ ++ /* If this fix's address is greater than the address of the first entry, ++ then we can't put the fix in this pool. We subtract the size of the ++ current fix to ensure that if the table is fully packed we still have ++ enough room to insert this value by suffling the other fixes forwards. */ ++ if (minipool_vector_head && ++ fix->address >= minipool_vector_head->max_address - fix->fix_size) ++ return NULL; ++ ++ /* Scan the pool to see if a constant with the same value has already been ++ added. While we are doing this, also note the location where we must ++ insert the constant if it doesn't already exist. */ ++ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) ++ { ++ if (GET_CODE (fix->value) == GET_CODE (mp->value) ++ && fix->mode == mp->mode ++ && (GET_CODE (fix->value) != CODE_LABEL ++ || (CODE_LABEL_NUMBER (fix->value) ++ == CODE_LABEL_NUMBER (mp->value))) ++ && rtx_equal_p (fix->value, mp->value)) ++ { ++ /* More than one fix references this entry. */ ++ mp->refcount++; ++ return move_minipool_fix_forward_ref (mp, max_mp, max_address); ++ } ++ ++ /* Note the insertion point if necessary. */ ++ if (max_mp == NULL && mp->max_address > max_address) ++ max_mp = mp; ++ ++ } ++ ++ /* The value is not currently in the minipool, so we need to create a new ++ entry for it. If MAX_MP is NULL, the entry will be put on the end of ++ the list since the placement is less constrained than any existing ++ entry. Otherwise, we insert the new fix before MAX_MP and, if ++ necessary, adjust the constraints on the other entries. */ ++ mp = xmalloc (sizeof (*mp)); ++ mp->fix_size = fix->fix_size; ++ mp->mode = fix->mode; ++ mp->value = fix->value; ++ mp->refcount = 1; ++ /* Not yet required for a backwards ref. */ ++ mp->min_address = -65536; ++ ++ if (max_mp == NULL) ++ { ++ mp->max_address = max_address; ++ mp->next = NULL; ++ mp->prev = minipool_vector_tail; ++ ++ if (mp->prev == NULL) ++ { ++ minipool_vector_head = mp; ++ minipool_vector_label = gen_label_rtx (); ++ } ++ else ++ mp->prev->next = mp; ++ ++ minipool_vector_tail = mp; ++ } ++ else ++ { ++ if (max_address > max_mp->max_address - mp->fix_size) ++ mp->max_address = max_mp->max_address - mp->fix_size; ++ else ++ mp->max_address = max_address; ++ ++ mp->next = max_mp; ++ mp->prev = max_mp->prev; ++ max_mp->prev = mp; ++ if (mp->prev != NULL) ++ mp->prev->next = mp; ++ else ++ minipool_vector_head = mp; ++ } ++ ++ /* Save the new entry. */ ++ max_mp = mp; ++ ++ /* Scan over the preceding entries and adjust their addresses as required. ++ */ ++ while (mp->prev != NULL ++ && mp->prev->max_address > mp->max_address - mp->prev->fix_size) ++ { ++ mp->prev->max_address = mp->max_address - mp->prev->fix_size; ++ mp = mp->prev; ++ } ++ ++ return max_mp; ++} ++ ++static Mnode * ++move_minipool_fix_backward_ref (Mnode * mp, Mnode * min_mp, ++ HOST_WIDE_INT min_address) ++ { ++ HOST_WIDE_INT offset; ++ ++ /* This should never be true, and the code below assumes these are ++ different. */ ++ if (mp == min_mp) ++ abort (); ++ ++ if (min_mp == NULL) ++ { ++ if (min_address > mp->min_address) ++ mp->min_address = min_address; ++ } ++ else ++ { ++ /* We will adjust this below if it is too loose. */ ++ mp->min_address = min_address; ++ ++ /* Unlink MP from its current position. Since min_mp is non-null, ++ mp->next must be non-null. */ ++ mp->next->prev = mp->prev; ++ if (mp->prev != NULL) ++ mp->prev->next = mp->next; ++ else ++ minipool_vector_head = mp->next; ++ ++ /* Reinsert it after MIN_MP. */ ++ mp->prev = min_mp; ++ mp->next = min_mp->next; ++ min_mp->next = mp; ++ if (mp->next != NULL) ++ mp->next->prev = mp; ++ else ++ minipool_vector_tail = mp; ++ } ++ ++ min_mp = mp; ++ ++ offset = 0; ++ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) ++ { ++ mp->offset = offset; ++ if (mp->refcount > 0) ++ offset += mp->fix_size; ++ ++ if (mp->next && mp->next->min_address < mp->min_address + mp->fix_size) ++ mp->next->min_address = mp->min_address + mp->fix_size; ++ } ++ ++ return min_mp; ++ } ++ ++/* Add a constant to the minipool for a backward reference. Returns the ++ node added or NULL if the constant will not fit in this pool. ++ ++ Note that the code for insertion for a backwards reference can be ++ somewhat confusing because the calculated offsets for each fix do ++ not take into account the size of the pool (which is still under ++ construction. */ ++static Mnode * ++add_minipool_backward_ref (Mfix * fix) ++{ ++ /* If set, min_mp is the last pool_entry that has a lower constraint than ++ the one we are trying to add. */ ++ Mnode *min_mp = NULL; ++ /* This can be negative, since it is only a constraint. */ ++ HOST_WIDE_INT min_address = fix->address - fix->backwards; ++ Mnode *mp; ++ ++ /* If we can't reach the current pool from this insn, or if we can't insert ++ this entry at the end of the pool without pushing other fixes out of ++ range, then we don't try. This ensures that we can't fail later on. */ ++ if (min_address >= minipool_barrier->address ++ || (minipool_vector_tail->min_address + fix->fix_size ++ >= minipool_barrier->address)) ++ return NULL; ++ ++ /* Scan the pool to see if a constant with the same value has already been ++ added. While we are doing this, also note the location where we must ++ insert the constant if it doesn't already exist. */ ++ for (mp = minipool_vector_tail; mp != NULL; mp = mp->prev) ++ { ++ if (GET_CODE (fix->value) == GET_CODE (mp->value) ++ && fix->mode == mp->mode ++ && (GET_CODE (fix->value) != CODE_LABEL ++ || (CODE_LABEL_NUMBER (fix->value) ++ == CODE_LABEL_NUMBER (mp->value))) ++ && rtx_equal_p (fix->value, mp->value) ++ /* Check that there is enough slack to move this entry to the end ++ of the table (this is conservative). */ ++ && (mp->max_address ++ > (minipool_barrier->address ++ + minipool_vector_tail->offset ++ + minipool_vector_tail->fix_size))) ++ { ++ mp->refcount++; ++ return move_minipool_fix_backward_ref (mp, min_mp, min_address); ++ } ++ ++ if (min_mp != NULL) ++ mp->min_address += fix->fix_size; ++ else ++ { ++ /* Note the insertion point if necessary. */ ++ if (mp->min_address < min_address) ++ { ++ min_mp = mp; ++ } ++ else if (mp->max_address ++ < minipool_barrier->address + mp->offset + fix->fix_size) ++ { ++ /* Inserting before this entry would push the fix beyond its ++ maximum address (which can happen if we have re-located a ++ forwards fix); force the new fix to come after it. */ ++ min_mp = mp; ++ min_address = mp->min_address + fix->fix_size; ++ } ++ } ++ } ++ ++ /* We need to create a new entry. */ ++ mp = xmalloc (sizeof (*mp)); ++ mp->fix_size = fix->fix_size; ++ mp->mode = fix->mode; ++ mp->value = fix->value; ++ mp->refcount = 1; ++ mp->max_address = minipool_barrier->address + 65536; ++ ++ mp->min_address = min_address; ++ ++ if (min_mp == NULL) ++ { ++ mp->prev = NULL; ++ mp->next = minipool_vector_head; ++ ++ if (mp->next == NULL) ++ { ++ minipool_vector_tail = mp; ++ minipool_vector_label = gen_label_rtx (); ++ } ++ else ++ mp->next->prev = mp; ++ ++ minipool_vector_head = mp; ++ } ++ else ++ { ++ mp->next = min_mp->next; ++ mp->prev = min_mp; ++ min_mp->next = mp; ++ ++ if (mp->next != NULL) ++ mp->next->prev = mp; ++ else ++ minipool_vector_tail = mp; ++ } ++ ++ /* Save the new entry. */ ++ min_mp = mp; ++ ++ if (mp->prev) ++ mp = mp->prev; ++ else ++ mp->offset = 0; ++ ++ /* Scan over the following entries and adjust their offsets. */ ++ while (mp->next != NULL) ++ { ++ if (mp->next->min_address < mp->min_address + mp->fix_size) ++ mp->next->min_address = mp->min_address + mp->fix_size; ++ ++ if (mp->refcount) ++ mp->next->offset = mp->offset + mp->fix_size; ++ else ++ mp->next->offset = mp->offset; ++ ++ mp = mp->next; ++ } ++ ++ return min_mp; ++} ++ ++static void ++assign_minipool_offsets (Mfix * barrier) ++ { ++ HOST_WIDE_INT offset = 0; ++ Mnode *mp; ++ ++ minipool_barrier = barrier; ++ ++ for (mp = minipool_vector_head; mp != NULL; mp = mp->next) ++ { ++ mp->offset = offset; ++ ++ if (mp->refcount > 0) ++ offset += mp->fix_size; ++ } ++ } ++ ++/* Print a symbolic form of X to the debug file, F. */ ++static void ++avr32_print_value (FILE * f, rtx x) ++ { ++ switch (GET_CODE (x)) ++ { ++ case CONST_INT: ++ fprintf (f, "0x%x", (int) INTVAL (x)); ++ return; ++ ++ case CONST_DOUBLE: ++ fprintf (f, "<0x%lx,0x%lx>", (long) XWINT (x, 2), (long) XWINT (x, 3)); ++ return; ++ ++ case CONST_VECTOR: ++ { ++ int i; ++ ++ fprintf (f, "<"); ++ for (i = 0; i < CONST_VECTOR_NUNITS (x); i++) ++ { ++ fprintf (f, "0x%x", (int) INTVAL (CONST_VECTOR_ELT (x, i))); ++ if (i < (CONST_VECTOR_NUNITS (x) - 1)) ++ fputc (',', f); ++ } ++ fprintf (f, ">"); ++ } ++ return; ++ ++ case CONST_STRING: ++ fprintf (f, "\"%s\"", XSTR (x, 0)); ++ return; ++ ++ case SYMBOL_REF: ++ fprintf (f, "`%s'", XSTR (x, 0)); ++ return; ++ ++ case LABEL_REF: ++ fprintf (f, "L%d", INSN_UID (XEXP (x, 0))); ++ return; ++ ++ case CONST: ++ avr32_print_value (f, XEXP (x, 0)); ++ return; ++ ++ case PLUS: ++ avr32_print_value (f, XEXP (x, 0)); ++ fprintf (f, "+"); ++ avr32_print_value (f, XEXP (x, 1)); ++ return; ++ ++ case PC: ++ fprintf (f, "pc"); ++ return; ++ ++ default: ++ fprintf (f, "????"); ++ return; ++ } ++ } ++ ++int ++is_minipool_label (rtx label) ++ { ++ minipool_labels *cur_mp_label = cfun->machine->minipool_label_head; ++ ++ if (GET_CODE (label) != CODE_LABEL) ++ return FALSE; ++ ++ while (cur_mp_label) ++ { ++ if (CODE_LABEL_NUMBER (label) ++ == CODE_LABEL_NUMBER (cur_mp_label->label)) ++ return TRUE; ++ cur_mp_label = cur_mp_label->next; ++ } ++ return FALSE; ++ } ++ ++static void ++new_minipool_label (rtx label) ++ { ++ if (!cfun->machine->minipool_label_head) ++ { ++ cfun->machine->minipool_label_head = ++ ggc_alloc (sizeof (minipool_labels)); ++ cfun->machine->minipool_label_tail = cfun->machine->minipool_label_head; ++ cfun->machine->minipool_label_head->label = label; ++ cfun->machine->minipool_label_head->next = 0; ++ cfun->machine->minipool_label_head->prev = 0; ++ } ++ else ++ { ++ cfun->machine->minipool_label_tail->next = ++ ggc_alloc (sizeof (minipool_labels)); ++ cfun->machine->minipool_label_tail->next->label = label; ++ cfun->machine->minipool_label_tail->next->next = 0; ++ cfun->machine->minipool_label_tail->next->prev = ++ cfun->machine->minipool_label_tail; ++ cfun->machine->minipool_label_tail = ++ cfun->machine->minipool_label_tail->next; ++ } ++ } ++ ++/* Output the literal table */ ++static void ++dump_minipool (rtx scan) ++ { ++ Mnode *mp; ++ Mnode *nmp; ++ ++ if (dump_file) ++ fprintf (dump_file, ++ ";; Emitting minipool after insn %u; address %ld; align %d (bytes)\n", ++ INSN_UID (scan), (unsigned long) minipool_barrier->address, 4); ++ ++ scan = emit_insn_after (gen_consttable_start (), scan); ++ scan = emit_insn_after (gen_align_4 (), scan); ++ scan = emit_label_after (minipool_vector_label, scan); ++ new_minipool_label (minipool_vector_label); ++ ++ for (mp = minipool_vector_head; mp != NULL; mp = nmp) ++ { ++ if (mp->refcount > 0) ++ { ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; Offset %u, min %ld, max %ld ", ++ (unsigned) mp->offset, (unsigned long) mp->min_address, ++ (unsigned long) mp->max_address); ++ avr32_print_value (dump_file, mp->value); ++ fputc ('\n', dump_file); ++ } ++ ++ switch (mp->fix_size) ++ { ++#ifdef HAVE_consttable_4 ++ case 4: ++ scan = emit_insn_after (gen_consttable_4 (mp->value), scan); ++ break; ++ ++#endif ++#ifdef HAVE_consttable_8 ++ case 8: ++ scan = emit_insn_after (gen_consttable_8 (mp->value), scan); ++ break; ++ ++#endif ++#ifdef HAVE_consttable_16 ++ case 16: ++ scan = emit_insn_after (gen_consttable_16 (mp->value), scan); ++ break; ++ ++#endif ++ case 0: ++ /* This can happen for force-minipool entries which just are ++ there to force the minipool to be generate. */ ++ break; ++ default: ++ abort (); ++ break; ++ } ++ } ++ ++ nmp = mp->next; ++ free (mp); ++ } ++ ++ minipool_vector_head = minipool_vector_tail = NULL; ++ scan = emit_insn_after (gen_consttable_end (), scan); ++ scan = emit_barrier_after (scan); ++ } ++ ++/* Return the cost of forcibly inserting a barrier after INSN. */ ++static int ++avr32_barrier_cost (rtx insn) ++ { ++ /* Basing the location of the pool on the loop depth is preferable, but at ++ the moment, the basic block information seems to be corrupt by this ++ stage of the compilation. */ ++ int base_cost = 50; ++ rtx next = next_nonnote_insn (insn); ++ ++ if (next != NULL && GET_CODE (next) == CODE_LABEL) ++ base_cost -= 20; ++ ++ switch (GET_CODE (insn)) ++ { ++ case CODE_LABEL: ++ /* It will always be better to place the table before the label, rather ++ than after it. */ ++ return 50; ++ ++ case INSN: ++ case CALL_INSN: ++ return base_cost; ++ ++ case JUMP_INSN: ++ return base_cost - 10; ++ ++ default: ++ return base_cost + 10; ++ } ++ } ++ ++/* Find the best place in the insn stream in the range ++ (FIX->address,MAX_ADDRESS) to forcibly insert a minipool barrier. ++ Create the barrier by inserting a jump and add a new fix entry for ++ it. */ ++static Mfix * ++create_fix_barrier (Mfix * fix, HOST_WIDE_INT max_address) ++{ ++ HOST_WIDE_INT count = 0; ++ rtx barrier; ++ rtx from = fix->insn; ++ rtx selected = from; ++ int selected_cost; ++ HOST_WIDE_INT selected_address; ++ Mfix *new_fix; ++ HOST_WIDE_INT max_count = max_address - fix->address; ++ rtx label = gen_label_rtx (); ++ ++ selected_cost = avr32_barrier_cost (from); ++ selected_address = fix->address; ++ ++ while (from && count < max_count) ++ { ++ rtx tmp; ++ int new_cost; ++ ++ /* This code shouldn't have been called if there was a natural barrier ++ within range. */ ++ if (GET_CODE (from) == BARRIER) ++ abort (); ++ ++ /* Count the length of this insn. */ ++ count += get_attr_length (from); ++ ++ /* If there is a jump table, add its length. */ ++ tmp = is_jump_table (from); ++ if (tmp != NULL) ++ { ++ count += get_jump_table_size (tmp); ++ ++ /* Jump tables aren't in a basic block, so base the cost on the ++ dispatch insn. If we select this location, we will still put ++ the pool after the table. */ ++ new_cost = avr32_barrier_cost (from); ++ ++ if (count < max_count && new_cost <= selected_cost) ++ { ++ selected = tmp; ++ selected_cost = new_cost; ++ selected_address = fix->address + count; ++ } ++ ++ /* Continue after the dispatch table. */ ++ from = NEXT_INSN (tmp); ++ continue; ++ } ++ ++ new_cost = avr32_barrier_cost (from); ++ ++ if (count < max_count && new_cost <= selected_cost) ++ { ++ selected = from; ++ selected_cost = new_cost; ++ selected_address = fix->address + count; ++ } ++ ++ from = NEXT_INSN (from); ++ } ++ ++ /* Create a new JUMP_INSN that branches around a barrier. */ ++ from = emit_jump_insn_after (gen_jump (label), selected); ++ JUMP_LABEL (from) = label; ++ barrier = emit_barrier_after (from); ++ emit_label_after (label, barrier); ++ ++ /* Create a minipool barrier entry for the new barrier. */ ++ new_fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*new_fix)); ++ new_fix->insn = barrier; ++ new_fix->address = selected_address; ++ new_fix->next = fix->next; ++ fix->next = new_fix; ++ ++ return new_fix; ++} ++ ++/* Record that there is a natural barrier in the insn stream at ++ ADDRESS. */ ++static void ++push_minipool_barrier (rtx insn, HOST_WIDE_INT address) ++ { ++ Mfix *fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*fix)); ++ ++ fix->insn = insn; ++ fix->address = address; ++ ++ fix->next = NULL; ++ if (minipool_fix_head != NULL) ++ minipool_fix_tail->next = fix; ++ else ++ minipool_fix_head = fix; ++ ++ minipool_fix_tail = fix; ++ } ++ ++/* Record INSN, which will need fixing up to load a value from the ++ minipool. ADDRESS is the offset of the insn since the start of the ++ function; LOC is a pointer to the part of the insn which requires ++ fixing; VALUE is the constant that must be loaded, which is of type ++ MODE. */ ++static void ++push_minipool_fix (rtx insn, HOST_WIDE_INT address, rtx * loc, ++ enum machine_mode mode, rtx value) ++ { ++ Mfix *fix = (Mfix *) obstack_alloc (&minipool_obstack, sizeof (*fix)); ++ rtx body = PATTERN (insn); ++ ++ fix->insn = insn; ++ fix->address = address; ++ fix->loc = loc; ++ fix->mode = mode; ++ fix->fix_size = MINIPOOL_FIX_SIZE (mode, value); ++ fix->value = value; ++ ++ if (GET_CODE (body) == PARALLEL) ++ { ++ /* Mcall : Ks16 << 2 */ ++ fix->forwards = ((1 << 15) - 1) << 2; ++ fix->backwards = (1 << 15) << 2; ++ } ++ else if (GET_CODE (body) == SET ++ && GET_MODE_SIZE (GET_MODE (SET_DEST (body))) == 4) ++ { ++ /* Word Load */ ++ if (TARGET_HARD_FLOAT ++ && GET_MODE_CLASS (GET_MODE (SET_DEST (body))) == MODE_FLOAT) ++ { ++ /* Ldc0.w : Ku12 << 2 */ ++ fix->forwards = ((1 << 12) - 1) << 2; ++ fix->backwards = 0; ++ } ++ else ++ { ++ if (optimize_size) ++ { ++ /* Lddpc : Ku7 << 2 */ ++ fix->forwards = ((1 << 7) - 1) << 2; ++ fix->backwards = 0; ++ } ++ else ++ { ++ /* Ld.w : Ks16 */ ++ fix->forwards = ((1 << 15) - 4); ++ fix->backwards = (1 << 15); ++ } ++ } ++ } ++ else if (GET_CODE (body) == SET ++ && GET_MODE_SIZE (GET_MODE (SET_DEST (body))) == 8) ++ { ++ /* Double word load */ ++ if (TARGET_HARD_FLOAT ++ && GET_MODE_CLASS (GET_MODE (SET_DEST (body))) == MODE_FLOAT) ++ { ++ /* Ldc0.d : Ku12 << 2 */ ++ fix->forwards = ((1 << 12) - 1) << 2; ++ fix->backwards = 0; ++ } ++ else ++ { ++ /* Ld.d : Ks16 */ ++ fix->forwards = ((1 << 15) - 4); ++ fix->backwards = (1 << 15); ++ } ++ } ++ else if (GET_CODE (body) == UNSPEC_VOLATILE ++ && XINT (body, 1) == VUNSPEC_MVRC) ++ { ++ /* Coprocessor load */ ++ /* Ldc : Ku8 << 2 */ ++ fix->forwards = ((1 << 8) - 1) << 2; ++ fix->backwards = 0; ++ } ++ else ++ { ++ /* Assume worst case which is lddpc insn. */ ++ fix->forwards = ((1 << 7) - 1) << 2; ++ fix->backwards = 0; ++ } ++ ++ fix->minipool = NULL; ++ ++ /* If an insn doesn't have a range defined for it, then it isn't expecting ++ to be reworked by this code. Better to abort now than to generate duff ++ assembly code. */ ++ if (fix->forwards == 0 && fix->backwards == 0) ++ abort (); ++ ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; %smode fixup for i%d; addr %lu, range (%ld,%ld): ", ++ GET_MODE_NAME (mode), ++ INSN_UID (insn), (unsigned long) address, ++ -1 * (long) fix->backwards, (long) fix->forwards); ++ avr32_print_value (dump_file, fix->value); ++ fprintf (dump_file, "\n"); ++ } ++ ++ /* Add it to the chain of fixes. */ ++ fix->next = NULL; ++ ++ if (minipool_fix_head != NULL) ++ minipool_fix_tail->next = fix; ++ else ++ minipool_fix_head = fix; ++ ++ minipool_fix_tail = fix; ++ } ++ ++/* Scan INSN and note any of its operands that need fixing. ++ If DO_PUSHES is false we do not actually push any of the fixups ++ needed. The function returns TRUE is any fixups were needed/pushed. ++ This is used by avr32_memory_load_p() which needs to know about loads ++ of constants that will be converted into minipool loads. */ ++static bool ++note_invalid_constants (rtx insn, HOST_WIDE_INT address, int do_pushes) ++ { ++ bool result = false; ++ int opno; ++ ++ extract_insn (insn); ++ ++ if (!constrain_operands (1)) ++ fatal_insn_not_found (insn); ++ ++ if (recog_data.n_alternatives == 0) ++ return false; ++ ++ /* Fill in recog_op_alt with information about the constraints of this ++ insn. */ ++ preprocess_constraints (); ++ ++ for (opno = 0; opno < recog_data.n_operands; opno++) ++ { ++ rtx op; ++ ++ /* Things we need to fix can only occur in inputs. */ ++ if (recog_data.operand_type[opno] != OP_IN) ++ continue; ++ ++ op = recog_data.operand[opno]; ++ ++ if (avr32_const_pool_ref_operand (op, GET_MODE (op))) ++ { ++ if (do_pushes) ++ { ++ rtx cop = avoid_constant_pool_reference (op); ++ ++ /* Casting the address of something to a mode narrower than a ++ word can cause avoid_constant_pool_reference() to return the ++ pool reference itself. That's no good to us here. Lets ++ just hope that we can use the constant pool value directly. ++ */ ++ if (op == cop) ++ cop = get_pool_constant (XEXP (op, 0)); ++ ++ push_minipool_fix (insn, address, ++ recog_data.operand_loc[opno], ++ recog_data.operand_mode[opno], cop); ++ } ++ ++ result = true; ++ } ++ else if (TARGET_HAS_ASM_ADDR_PSEUDOS ++ && avr32_address_operand (op, GET_MODE (op))) ++ { ++ /* Handle pseudo instructions using a direct address. These pseudo ++ instructions might need entries in the constant pool and we must ++ therefor create a constant pool for them, in case the ++ assembler/linker needs to insert entries. */ ++ if (do_pushes) ++ { ++ /* Push a dummy constant pool entry so that the .cpool ++ directive should be inserted on the appropriate place in the ++ code even if there are no real constant pool entries. This ++ is used by the assembler and linker to know where to put ++ generated constant pool entries. */ ++ push_minipool_fix (insn, address, ++ recog_data.operand_loc[opno], ++ recog_data.operand_mode[opno], ++ gen_rtx_UNSPEC (VOIDmode, ++ gen_rtvec (1, const0_rtx), ++ UNSPEC_FORCE_MINIPOOL)); ++ result = true; ++ } ++ } ++ } ++ return result; ++ } ++ ++ ++static int ++avr32_insn_is_cast (rtx insn) ++ { ++ ++ if (NONJUMP_INSN_P (insn) ++ && GET_CODE (PATTERN (insn)) == SET ++ && (GET_CODE (SET_SRC (PATTERN (insn))) == ZERO_EXTEND ++ || GET_CODE (SET_SRC (PATTERN (insn))) == SIGN_EXTEND) ++ && REG_P (XEXP (SET_SRC (PATTERN (insn)), 0)) ++ && REG_P (SET_DEST (PATTERN (insn)))) ++ return true; ++ return false; ++ } ++ ++/* FIXME: The level of nesting in this function is way too deep. It needs to be ++ torn apart. */ ++static void ++avr32_reorg_optimization (void) ++ { ++ rtx first = get_insns (); ++ rtx insn; ++ ++ if (TARGET_MD_REORG_OPTIMIZATION && (optimize_size || (optimize > 0))) ++ { ++ ++ /* Scan through all insns looking for cast operations. */ ++ if (dump_file) ++ { ++ fprintf (dump_file, ";; Deleting redundant cast operations:\n"); ++ } ++ for (insn = first; insn; insn = NEXT_INSN (insn)) ++ { ++ rtx reg, src_reg, scan; ++ enum machine_mode mode; ++ int unused_cast; ++ rtx label_ref; ++ ++ if (avr32_insn_is_cast (insn) ++ && (GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == QImode ++ || GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == HImode)) ++ { ++ mode = GET_MODE (XEXP (SET_SRC (PATTERN (insn)), 0)); ++ reg = SET_DEST (PATTERN (insn)); ++ src_reg = XEXP (SET_SRC (PATTERN (insn)), 0); ++ } ++ else ++ { ++ continue; ++ } ++ ++ unused_cast = false; ++ label_ref = NULL_RTX; ++ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) ++ { ++ /* Check if we have reached the destination of a simple ++ conditional jump which we have already scanned past. If so, ++ we can safely continue scanning. */ ++ if (LABEL_P (scan) && label_ref != NULL_RTX) ++ { ++ if (CODE_LABEL_NUMBER (scan) == ++ CODE_LABEL_NUMBER (XEXP (label_ref, 0))) ++ label_ref = NULL_RTX; ++ else ++ break; ++ } ++ ++ if (!INSN_P (scan)) ++ continue; ++ ++ /* For conditional jumps we can manage to keep on scanning if ++ we meet the destination label later on before any new jump ++ insns occure. */ ++ if (GET_CODE (scan) == JUMP_INSN) ++ { ++ if (any_condjump_p (scan) && label_ref == NULL_RTX) ++ label_ref = condjump_label (scan); ++ else ++ break; ++ } ++ ++ if (!reg_mentioned_p (reg, PATTERN (scan))) ++ continue; ++ ++ /* Check if casted register is used in this insn */ ++ if ((regno_use_in (REGNO (reg), PATTERN (scan)) != NULL_RTX) ++ && (GET_MODE (regno_use_in (REGNO (reg), PATTERN (scan))) == ++ GET_MODE (reg))) ++ { ++ /* If not used in the source to the set or in a memory ++ expression in the destiantion then the register is used ++ as a destination and is really dead. */ ++ if (single_set (scan) ++ && GET_CODE (PATTERN (scan)) == SET ++ && REG_P (SET_DEST (PATTERN (scan))) ++ && !regno_use_in (REGNO (reg), SET_SRC (PATTERN (scan))) ++ && label_ref == NULL_RTX) ++ { ++ unused_cast = true; ++ } ++ break; ++ } ++ ++ /* Check if register is dead or set in this insn */ ++ if (dead_or_set_p (scan, reg)) ++ { ++ unused_cast = true; ++ break; ++ } ++ } ++ ++ /* Check if we have unresolved conditional jumps */ ++ if (label_ref != NULL_RTX) ++ continue; ++ ++ if (unused_cast) ++ { ++ if (REGNO (reg) == REGNO (XEXP (SET_SRC (PATTERN (insn)), 0))) ++ { ++ /* One operand cast, safe to delete */ ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; INSN %i removed, casted register %i value not used.\n", ++ INSN_UID (insn), REGNO (reg)); ++ } ++ SET_INSN_DELETED (insn); ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (insn) = -1; ++ } ++ else ++ { ++ /* Two operand cast, which really could be substituted with ++ a move, if the source register is dead after the cast ++ insn and then the insn which sets the source register ++ could instead directly set the destination register for ++ the cast. As long as there are no insns in between which ++ uses the register. */ ++ rtx link = NULL_RTX; ++ rtx set; ++ rtx src_reg = XEXP (SET_SRC (PATTERN (insn)), 0); ++ unused_cast = false; ++ ++ if (!find_reg_note (insn, REG_DEAD, src_reg)) ++ continue; ++ ++ /* Search for the insn which sets the source register */ ++ for (link = LOG_LINKS (insn); link; link = XEXP (link, 1)) ++ { ++ if (REG_NOTE_KIND (link) != 0) ++ continue; ++ set = single_set (XEXP (link, 0)); ++ if (set && rtx_equal_p (src_reg, SET_DEST (set))) ++ { ++ link = XEXP (link, 0); ++ break; ++ } ++ } ++ ++ /* Found no link or link is a call insn where we can not ++ change the destination register */ ++ if (link == NULL_RTX || CALL_P (link)) ++ continue; ++ ++ /* Scan through all insn between link and insn */ ++ for (scan = NEXT_INSN (link); scan; scan = NEXT_INSN (scan)) ++ { ++ /* Don't try to trace forward past a CODE_LABEL if we ++ haven't seen INSN yet. Ordinarily, we will only ++ find the setting insn in LOG_LINKS if it is in the ++ same basic block. However, cross-jumping can insert ++ code labels in between the load and the call, and ++ can result in situations where a single call insn ++ may have two targets depending on where we came ++ from. */ ++ ++ if (GET_CODE (scan) == CODE_LABEL) ++ break; ++ ++ if (!INSN_P (scan)) ++ continue; ++ ++ /* Don't try to trace forward past a JUMP. To optimize ++ safely, we would have to check that all the ++ instructions at the jump destination did not use REG. ++ */ ++ ++ if (GET_CODE (scan) == JUMP_INSN) ++ { ++ break; ++ } ++ ++ if (!reg_mentioned_p (src_reg, PATTERN (scan))) ++ continue; ++ ++ /* We have reached the cast insn */ ++ if (scan == insn) ++ { ++ /* We can remove cast and replace the destination ++ register of the link insn with the destination ++ of the cast */ ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; INSN %i removed, casted value unused. " ++ "Destination of removed cast operation: register %i, folded into INSN %i.\n", ++ INSN_UID (insn), REGNO (reg), ++ INSN_UID (link)); ++ } ++ /* Update link insn */ ++ SET_DEST (PATTERN (link)) = ++ gen_rtx_REG (mode, REGNO (reg)); ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (link) = -1; ++ ++ /* Delete insn */ ++ SET_INSN_DELETED (insn); ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (insn) = -1; ++ break; ++ } ++ } ++ } ++ } ++ } ++ } ++ ++ if (TARGET_MD_REORG_OPTIMIZATION && (optimize_size || (optimize > 0))) ++ { ++ ++ /* Scan through all insns looking for shifted add operations */ ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; Deleting redundant shifted add operations:\n"); ++ } ++ for (insn = first; insn; insn = NEXT_INSN (insn)) ++ { ++ rtx reg, mem_expr, scan, op0, op1; ++ int add_only_used_as_pointer; ++ ++ if (INSN_P (insn) ++ && GET_CODE (PATTERN (insn)) == SET ++ && GET_CODE (SET_SRC (PATTERN (insn))) == PLUS ++ && (GET_CODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == MULT ++ || GET_CODE (XEXP (SET_SRC (PATTERN (insn)), 0)) == ASHIFT) ++ && GET_CODE (XEXP (XEXP (SET_SRC (PATTERN (insn)), 0), 1)) == ++ CONST_INT && REG_P (SET_DEST (PATTERN (insn))) ++ && REG_P (XEXP (SET_SRC (PATTERN (insn)), 1)) ++ && REG_P (XEXP (XEXP (SET_SRC (PATTERN (insn)), 0), 0))) ++ { ++ reg = SET_DEST (PATTERN (insn)); ++ mem_expr = SET_SRC (PATTERN (insn)); ++ op0 = XEXP (XEXP (mem_expr, 0), 0); ++ op1 = XEXP (mem_expr, 1); ++ } ++ else ++ { ++ continue; ++ } ++ ++ /* Scan forward the check if the result of the shifted add ++ operation is only used as an address in memory operations and ++ that the operands to the shifted add are not clobbered. */ ++ add_only_used_as_pointer = false; ++ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) ++ { ++ if (!INSN_P (scan)) ++ continue; ++ ++ /* Don't try to trace forward past a JUMP or CALL. To optimize ++ safely, we would have to check that all the instructions at ++ the jump destination did not use REG. */ ++ ++ if (GET_CODE (scan) == JUMP_INSN) ++ { ++ break; ++ } ++ ++ /* If used in a call insn then we cannot optimize it away */ ++ if (CALL_P (scan) && find_regno_fusage (scan, USE, REGNO (reg))) ++ break; ++ ++ /* If any of the operands of the shifted add are clobbered we ++ cannot optimize the shifted adda away */ ++ if ((reg_set_p (op0, scan) && (REGNO (op0) != REGNO (reg))) ++ || (reg_set_p (op1, scan) && (REGNO (op1) != REGNO (reg)))) ++ break; ++ ++ if (!reg_mentioned_p (reg, PATTERN (scan))) ++ continue; ++ ++ /* If used any other place than as a pointer or as the ++ destination register we failed */ ++ if (!(single_set (scan) ++ && GET_CODE (PATTERN (scan)) == SET ++ && ((MEM_P (SET_DEST (PATTERN (scan))) ++ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) ++ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == ++ REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) ++ && ++ REG_P (XEXP ++ (SET_SRC (PATTERN (scan)), ++ 0)) ++ && ++ REGNO (XEXP ++ (SET_SRC (PATTERN (scan)), ++ 0)) == REGNO (reg)))) ++ && !(GET_CODE (PATTERN (scan)) == SET ++ && REG_P (SET_DEST (PATTERN (scan))) ++ && !regno_use_in (REGNO (reg), ++ SET_SRC (PATTERN (scan))))) ++ break; ++ ++ /* Check if register is dead or set in this insn */ ++ if (dead_or_set_p (scan, reg)) ++ { ++ add_only_used_as_pointer = true; ++ break; ++ } ++ } ++ ++ if (add_only_used_as_pointer) ++ { ++ /* Lets delete the add insn and replace all memory references ++ which uses the pointer with the full expression. */ ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; Deleting INSN %i since address expression can be folded into all " ++ "memory references using this expression\n", ++ INSN_UID (insn)); ++ } ++ SET_INSN_DELETED (insn); ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (insn) = -1; ++ ++ for (scan = NEXT_INSN (insn); scan; scan = NEXT_INSN (scan)) ++ { ++ if (!INSN_P (scan)) ++ continue; ++ ++ if (!reg_mentioned_p (reg, PATTERN (scan))) ++ continue; ++ ++ /* If used any other place than as a pointer or as the ++ destination register we failed */ ++ if ((single_set (scan) ++ && GET_CODE (PATTERN (scan)) == SET ++ && ((MEM_P (SET_DEST (PATTERN (scan))) ++ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) ++ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == ++ REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) ++ && ++ REG_P (XEXP ++ (SET_SRC (PATTERN (scan)), ++ 0)) ++ && ++ REGNO (XEXP ++ (SET_SRC (PATTERN (scan)), ++ 0)) == REGNO (reg))))) ++ { ++ if (dump_file) ++ { ++ fprintf (dump_file, ++ ";; Register %i replaced by indexed address in INSN %i\n", ++ REGNO (reg), INSN_UID (scan)); ++ } ++ if (MEM_P (SET_DEST (PATTERN (scan)))) ++ XEXP (SET_DEST (PATTERN (scan)), 0) = mem_expr; ++ else ++ XEXP (SET_SRC (PATTERN (scan)), 0) = mem_expr; ++ } ++ ++ /* Check if register is dead or set in this insn */ ++ if (dead_or_set_p (scan, reg)) ++ { ++ break; ++ } ++ ++ } ++ } ++ } ++ } ++ } ++ ++/* Exported to toplev.c. ++ ++ Do a final pass over the function, just before delayed branch ++ scheduling. */ ++ ++static void ++avr32_reorg (void) ++ { ++ rtx insn; ++ HOST_WIDE_INT address = 0; ++ Mfix *fix; ++ ++ minipool_fix_head = minipool_fix_tail = NULL; ++ ++ /* The first insn must always be a note, or the code below won't scan it ++ properly. */ ++ insn = get_insns (); ++ if (GET_CODE (insn) != NOTE) ++ abort (); ++ ++ /* Scan all the insns and record the operands that will need fixing. */ ++ for (insn = next_nonnote_insn (insn); insn; insn = next_nonnote_insn (insn)) ++ { ++ if (GET_CODE (insn) == BARRIER) ++ push_minipool_barrier (insn, address); ++ else if (INSN_P (insn)) ++ { ++ rtx table; ++ ++ note_invalid_constants (insn, address, true); ++ address += get_attr_length (insn); ++ ++ /* If the insn is a vector jump, add the size of the table and skip ++ the table. */ ++ if ((table = is_jump_table (insn)) != NULL) ++ { ++ address += get_jump_table_size (table); ++ insn = table; ++ } ++ } ++ } ++ ++ fix = minipool_fix_head; ++ ++ /* Now scan the fixups and perform the required changes. */ ++ while (fix) ++ { ++ Mfix *ftmp; ++ Mfix *fdel; ++ Mfix *last_added_fix; ++ Mfix *last_barrier = NULL; ++ Mfix *this_fix; ++ ++ /* Skip any further barriers before the next fix. */ ++ while (fix && GET_CODE (fix->insn) == BARRIER) ++ fix = fix->next; ++ ++ /* No more fixes. */ ++ if (fix == NULL) ++ break; ++ ++ last_added_fix = NULL; ++ ++ for (ftmp = fix; ftmp; ftmp = ftmp->next) ++ { ++ if (GET_CODE (ftmp->insn) == BARRIER) ++ { ++ if (ftmp->address >= minipool_vector_head->max_address) ++ break; ++ ++ last_barrier = ftmp; ++ } ++ else if ((ftmp->minipool = add_minipool_forward_ref (ftmp)) == NULL) ++ break; ++ ++ last_added_fix = ftmp; /* Keep track of the last fix added. ++ */ ++ } ++ ++ /* If we found a barrier, drop back to that; any fixes that we could ++ have reached but come after the barrier will now go in the next ++ mini-pool. */ ++ if (last_barrier != NULL) ++ { ++ /* Reduce the refcount for those fixes that won't go into this pool ++ after all. */ ++ for (fdel = last_barrier->next; ++ fdel && fdel != ftmp; fdel = fdel->next) ++ { ++ fdel->minipool->refcount--; ++ fdel->minipool = NULL; ++ } ++ ++ ftmp = last_barrier; ++ } ++ else ++ { ++ /* ftmp is first fix that we can't fit into this pool and there no ++ natural barriers that we could use. Insert a new barrier in the ++ code somewhere between the previous fix and this one, and ++ arrange to jump around it. */ ++ HOST_WIDE_INT max_address; ++ ++ /* The last item on the list of fixes must be a barrier, so we can ++ never run off the end of the list of fixes without last_barrier ++ being set. */ ++ if (ftmp == NULL) ++ abort (); ++ ++ max_address = minipool_vector_head->max_address; ++ /* Check that there isn't another fix that is in range that we ++ couldn't fit into this pool because the pool was already too ++ large: we need to put the pool before such an instruction. */ ++ if (ftmp->address < max_address) ++ max_address = ftmp->address; ++ ++ last_barrier = create_fix_barrier (last_added_fix, max_address); ++ } ++ ++ assign_minipool_offsets (last_barrier); ++ ++ while (ftmp) ++ { ++ if (GET_CODE (ftmp->insn) != BARRIER ++ && ((ftmp->minipool = add_minipool_backward_ref (ftmp)) ++ == NULL)) ++ break; ++ ++ ftmp = ftmp->next; ++ } ++ ++ /* Scan over the fixes we have identified for this pool, fixing them up ++ and adding the constants to the pool itself. */ ++ for (this_fix = fix; this_fix && ftmp != this_fix; ++ this_fix = this_fix->next) ++ if (GET_CODE (this_fix->insn) != BARRIER ++ /* Do nothing for entries present just to force the insertion of ++ a minipool. */ ++ && !IS_FORCE_MINIPOOL (this_fix->value)) ++ { ++ rtx addr = plus_constant (gen_rtx_LABEL_REF (VOIDmode, ++ minipool_vector_label), ++ this_fix->minipool->offset); ++ *this_fix->loc = gen_rtx_MEM (this_fix->mode, addr); ++ } ++ ++ dump_minipool (last_barrier->insn); ++ fix = ftmp; ++ } ++ ++ /* Free the minipool memory. */ ++ obstack_free (&minipool_obstack, minipool_startobj); ++ ++ avr32_reorg_optimization (); ++ } ++ ++ ++/* ++ Hook for doing some final scanning of instructions. Does nothing yet...*/ ++void ++avr32_final_prescan_insn (rtx insn ATTRIBUTE_UNUSED, ++ rtx * opvec ATTRIBUTE_UNUSED, ++ int noperands ATTRIBUTE_UNUSED) ++ { ++ return; ++ } ++ ++ ++/* Function for changing the condition on the next instruction, ++ should be used when emmiting compare instructions and ++ the condition of the next instruction needs to change. ++ */ ++int ++set_next_insn_cond (rtx cur_insn, rtx new_cond) ++ { ++ rtx next_insn = next_nonnote_insn (cur_insn); ++ if ((next_insn != NULL_RTX) ++ && (INSN_P (next_insn)) ++ && (GET_CODE (PATTERN (next_insn)) == SET) ++ && (GET_CODE (SET_SRC (PATTERN (next_insn))) == IF_THEN_ELSE)) ++ { ++ /* Branch instructions */ ++ XEXP (SET_SRC (PATTERN (next_insn)), 0) = new_cond; ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (next_insn) = -1; ++ return TRUE; ++ } ++ else if ((next_insn != NULL_RTX) ++ && (INSN_P (next_insn)) ++ && (GET_CODE (PATTERN (next_insn)) == SET) ++ && comparison_operator (SET_SRC (PATTERN (next_insn)), ++ GET_MODE (SET_SRC (PATTERN (next_insn))))) ++ { ++ /* scc with no compare */ ++ SET_SRC (PATTERN (next_insn)) = new_cond; ++ /* Force the instruction to be recognized again */ ++ INSN_CODE (next_insn) = -1; ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++/* Function for obtaining the condition for the next instruction ++ after cur_insn. ++ */ ++rtx ++get_next_insn_cond (rtx cur_insn) ++ { ++ rtx next_insn = next_nonnote_insn (cur_insn); ++ rtx cond = NULL_RTX; ++ if ((next_insn != NULL_RTX) ++ && (INSN_P (next_insn)) ++ && (GET_CODE (PATTERN (next_insn)) == SET) ++ && (GET_CODE (SET_SRC (PATTERN (next_insn))) == IF_THEN_ELSE)) ++ { ++ /* Branch instructions */ ++ cond = XEXP (SET_SRC (PATTERN (next_insn)), 0); ++ } ++ else if ((next_insn != NULL_RTX) ++ && (INSN_P (next_insn)) ++ && (GET_CODE (PATTERN (next_insn)) == SET) ++ && comparison_operator (SET_SRC (PATTERN (next_insn)), ++ GET_MODE (SET_SRC (PATTERN (next_insn))))) ++ { ++ /* scc with no compare */ ++ cond = SET_SRC (PATTERN (next_insn)); ++ } ++ ++ return cond; ++ } ++ ++ ++rtx ++avr32_output_cmp (rtx cond, enum machine_mode mode, rtx op0, rtx op1) ++ { ++ ++ rtx new_cond = NULL_RTX; ++ rtx ops[2]; ++ rtx compare_pattern; ++ ops[0] = op0; ++ ops[1] = op1; ++ ++ compare_pattern = gen_rtx_COMPARE (mode, op0, op1); ++ ++ new_cond = is_compare_redundant (compare_pattern, cond); ++ ++ if (new_cond != NULL_RTX) ++ return new_cond; ++ ++ /* Insert compare */ ++ switch (mode) ++ { ++ case QImode: ++ output_asm_insn ("cp.b\t%0, %1", ops); ++ break; ++ case HImode: ++ output_asm_insn ("cp.h\t%0, %1", ops); ++ break; ++ case SImode: ++ output_asm_insn ("cp.w\t%0, %1", ops); ++ break; ++ case DImode: ++ if (GET_CODE (op1) != REG) ++ output_asm_insn ("cp.w\t%0, %1\ncpc\t%m0", ops); ++ else ++ output_asm_insn ("cp.w\t%0, %1\ncpc\t%m0, %m1", ops); ++ break; ++ default: ++ internal_error ("Unknown comparison mode"); ++ break; ++ } ++ ++ return cond; ++ } ++ ++int ++avr32_load_multiple_operation (rtx op, ++ enum machine_mode mode ATTRIBUTE_UNUSED) ++ { ++ int count = XVECLEN (op, 0); ++ unsigned int dest_regno; ++ rtx src_addr; ++ rtx elt; ++ int i = 1, base = 0; ++ ++ if (count <= 1 || GET_CODE (XVECEXP (op, 0, 0)) != SET) ++ return 0; ++ ++ /* Check to see if this might be a write-back. */ ++ if (GET_CODE (SET_SRC (elt = XVECEXP (op, 0, 0))) == PLUS) ++ { ++ i++; ++ base = 1; ++ ++ /* Now check it more carefully. */ ++ if (GET_CODE (SET_DEST (elt)) != REG ++ || GET_CODE (XEXP (SET_SRC (elt), 0)) != REG ++ || GET_CODE (XEXP (SET_SRC (elt), 1)) != CONST_INT ++ || INTVAL (XEXP (SET_SRC (elt), 1)) != (count - 1) * 4) ++ return 0; ++ } ++ ++ /* Perform a quick check so we don't blow up below. */ ++ if (count <= 1 ++ || GET_CODE (XVECEXP (op, 0, i - 1)) != SET ++ || GET_CODE (SET_DEST (XVECEXP (op, 0, i - 1))) != REG ++ || GET_CODE (SET_SRC (XVECEXP (op, 0, i - 1))) != UNSPEC) ++ return 0; ++ ++ dest_regno = REGNO (SET_DEST (XVECEXP (op, 0, i - 1))); ++ src_addr = XEXP (SET_SRC (XVECEXP (op, 0, i - 1)), 0); ++ ++ for (; i < count; i++) ++ { ++ elt = XVECEXP (op, 0, i); ++ ++ if (GET_CODE (elt) != SET ++ || GET_CODE (SET_DEST (elt)) != REG ++ || GET_MODE (SET_DEST (elt)) != SImode ++ || GET_CODE (SET_SRC (elt)) != UNSPEC) ++ return 0; ++ } ++ ++ return 1; ++ } ++ ++int ++avr32_store_multiple_operation (rtx op, ++ enum machine_mode mode ATTRIBUTE_UNUSED) ++ { ++ int count = XVECLEN (op, 0); ++ int src_regno; ++ rtx dest_addr; ++ rtx elt; ++ int i = 1; ++ ++ if (count <= 1 || GET_CODE (XVECEXP (op, 0, 0)) != SET) ++ return 0; ++ ++ /* Perform a quick check so we don't blow up below. */ ++ if (count <= i ++ || GET_CODE (XVECEXP (op, 0, i - 1)) != SET ++ || GET_CODE (SET_DEST (XVECEXP (op, 0, i - 1))) != MEM ++ || GET_CODE (SET_SRC (XVECEXP (op, 0, i - 1))) != UNSPEC) ++ return 0; ++ ++ src_regno = REGNO (SET_SRC (XVECEXP (op, 0, i - 1))); ++ dest_addr = XEXP (SET_DEST (XVECEXP (op, 0, i - 1)), 0); ++ ++ for (; i < count; i++) ++ { ++ elt = XVECEXP (op, 0, i); ++ ++ if (GET_CODE (elt) != SET ++ || GET_CODE (SET_DEST (elt)) != MEM ++ || GET_MODE (SET_DEST (elt)) != SImode ++ || GET_CODE (SET_SRC (elt)) != UNSPEC) ++ return 0; ++ } ++ ++ return 1; ++ } ++ ++int ++avr32_valid_macmac_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* Check if they use the same accumulator */ ++ if (rtx_equal_p ++ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) ++ { ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++int ++avr32_valid_mulmac_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* ++ Check if the mul instruction produces the accumulator for the mac ++ instruction. */ ++ if (rtx_equal_p ++ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) ++ { ++ return TRUE; ++ } ++ return FALSE; ++ } ++ ++int ++avr32_store_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* Only valid bypass if the output result is used as an src in the store ++ instruction, NOT if used as a pointer or base. */ ++ if (rtx_equal_p ++ (SET_DEST (PATTERN (insn_out)), SET_SRC (PATTERN (insn_in)))) ++ { ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++int ++avr32_mul_waw_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* Check if the register holding the result from the mul instruction is ++ used as a result register in the input instruction. */ ++ if (rtx_equal_p ++ (SET_DEST (PATTERN (insn_out)), SET_DEST (PATTERN (insn_in)))) ++ { ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++int ++avr32_valid_load_double_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* Check if the first loaded word in insn_out is used in insn_in. */ ++ rtx dst_reg; ++ rtx second_loaded_reg; ++ ++ /* If this is a double alu operation then the bypass is not valid */ ++ if ((get_attr_type (insn_in) == TYPE_ALU ++ || get_attr_type (insn_in) == TYPE_ALU2) ++ && (GET_MODE_SIZE (GET_MODE (SET_DEST (PATTERN (insn_out)))) > 4)) ++ return FALSE; ++ ++ /* Get the destination register in the load */ ++ if (!REG_P (SET_DEST (PATTERN (insn_out)))) ++ return FALSE; ++ ++ dst_reg = SET_DEST (PATTERN (insn_out)); ++ second_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 1); ++ ++ if (!reg_mentioned_p (second_loaded_reg, PATTERN (insn_in))) ++ return TRUE; ++ ++ return FALSE; ++ } ++ ++ ++int ++avr32_valid_load_quad_bypass (rtx insn_out, rtx insn_in) ++ { ++ /* ++ Check if the two first loaded word in insn_out are used in insn_in. */ ++ rtx dst_reg; ++ rtx third_loaded_reg, fourth_loaded_reg; ++ ++ /* Get the destination register in the load */ ++ if (!REG_P (SET_DEST (PATTERN (insn_out)))) ++ return FALSE; ++ ++ dst_reg = SET_DEST (PATTERN (insn_out)); ++ third_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 2); ++ fourth_loaded_reg = gen_rtx_REG (SImode, REGNO (dst_reg) + 3); ++ ++ if (!reg_mentioned_p (third_loaded_reg, PATTERN (insn_in)) ++ && !reg_mentioned_p (fourth_loaded_reg, PATTERN (insn_in))) ++ { ++ return TRUE; ++ } ++ ++ return FALSE; ++ } ++ ++ ++//section * ++//avr32_select_rtx_section (enum machine_mode mode ATTRIBUTE_UNUSED, ++// rtx x ATTRIBUTE_UNUSED, ++// unsigned HOST_WIDE_INT align ATTRIBUTE_UNUSED) ++// { ++// /* Let ASM_OUTPUT_POOL_PROLOGUE take care of this */ ++// return 0; ++// } ++ ++ ++/* Function for getting an integer value from a const_int or const_double ++ expression regardless of the HOST_WIDE_INT size. Each target cpu word ++ will be put into the val array where the LSW will be stored at the lowest ++ address and so forth. Assumes that const_expr is either a const_int or ++ const_double. Only valid for modes which have sizes that are a multiple ++ of the word size. ++*/ ++void ++avr32_get_intval (enum machine_mode mode, ++ rtx const_expr, ++ HOST_WIDE_INT *val) ++{ ++ int words_in_mode = GET_MODE_SIZE (mode)/UNITS_PER_WORD; ++ const int words_in_const_int = HOST_BITS_PER_WIDE_INT / BITS_PER_WORD; ++ ++ if ( GET_CODE(const_expr) == CONST_DOUBLE ){ ++ HOST_WIDE_INT hi = CONST_DOUBLE_HIGH(const_expr); ++ HOST_WIDE_INT lo = CONST_DOUBLE_LOW(const_expr); ++ /* Evaluate hi and lo values of const_double. */ ++ avr32_get_intval (mode_for_size (HOST_BITS_PER_WIDE_INT, MODE_INT, 0), ++ GEN_INT (lo), ++ &val[0]); ++ avr32_get_intval (mode_for_size (HOST_BITS_PER_WIDE_INT, MODE_INT, 0), ++ GEN_INT (hi), ++ &val[words_in_const_int]); ++ } else if ( GET_CODE(const_expr) == CONST_INT ){ ++ HOST_WIDE_INT value = INTVAL(const_expr); ++ int word; ++ for ( word = 0; (word < words_in_mode) && (word < words_in_const_int); word++ ){ ++ /* Shift word up to the MSW and shift down again to extract the ++ word and sign-extend. */ ++ int lshift = (words_in_const_int - word - 1) * BITS_PER_WORD; ++ int rshift = (words_in_const_int-1) * BITS_PER_WORD; ++ val[word] = (value << lshift) >> rshift; ++ } ++ ++ for ( ; word < words_in_mode; word++ ){ ++ /* Just put the sign bits in the remaining words. */ ++ val[word] = value < 0 ? -1 : 0; ++ } ++ } ++} ++ ++void ++avr32_split_const_expr (enum machine_mode mode, ++ enum machine_mode new_mode, ++ rtx expr, ++ rtx *split_expr) ++{ ++ int i, word; ++ int words_in_intval = GET_MODE_SIZE (mode)/UNITS_PER_WORD; ++ int words_in_split_values = GET_MODE_SIZE (new_mode)/UNITS_PER_WORD; ++ const int words_in_const_int = HOST_BITS_PER_WIDE_INT / BITS_PER_WORD; ++ HOST_WIDE_INT *val = alloca (words_in_intval * UNITS_PER_WORD); ++ ++ avr32_get_intval (mode, expr, val); ++ ++ for ( i=0; i < (words_in_intval/words_in_split_values); i++ ) ++ { ++ HOST_WIDE_INT value_lo = 0, value_hi = 0; ++ for ( word = 0; word < words_in_split_values; word++ ) ++ { ++ if ( word >= words_in_const_int ) ++ value_hi |= ((val[i * words_in_split_values + word] & ++ (((HOST_WIDE_INT)1 << BITS_PER_WORD)-1)) ++ << (BITS_PER_WORD * (word - words_in_const_int))); ++ else ++ value_lo |= ((val[i * words_in_split_values + word] & ++ (((HOST_WIDE_INT)1 << BITS_PER_WORD)-1)) ++ << (BITS_PER_WORD * word)); ++ } ++ split_expr[i] = immed_double_const(value_lo, value_hi, new_mode); ++ } ++} ++ ++ ++ ++/* Set up library functions to comply to AVR32 ABI */ ++ ++ ++/* Set up library functions to comply to AVR32 ABI */ ++ ++static void ++avr32_init_libfuncs (void) ++{ ++ /* Convert gcc run-time function names to AVR32 ABI names */ ++ ++ /* Double-precision floating-point arithmetic. */ ++ set_optab_libfunc (sdiv_optab, DFmode, "__avr32_f64_div"); ++ set_optab_libfunc (smul_optab, DFmode, "__avr32_f64_mul"); ++ set_optab_libfunc (neg_optab, DFmode, NULL); ++ ++ /* Double-precision comparisons. */ ++ set_optab_libfunc (eq_optab, DFmode, "__avr32_f64_cmp_eq"); ++ set_optab_libfunc (ne_optab, DFmode, NULL); ++ set_optab_libfunc (lt_optab, DFmode, "__avr32_f64_cmp_lt"); ++ set_optab_libfunc (le_optab, DFmode, NULL); ++ set_optab_libfunc (ge_optab, DFmode, "__avr32_f64_cmp_ge"); ++ set_optab_libfunc (gt_optab, DFmode, NULL); ++ ++ /* Single-precision floating-point arithmetic. */ ++ set_optab_libfunc (smul_optab, SFmode, "__avr32_f32_mul"); ++ set_optab_libfunc (neg_optab, SFmode, NULL); ++ ++ /* Single-precision comparisons. */ ++ set_optab_libfunc (eq_optab, SFmode, "__avr32_f32_cmp_eq"); ++ set_optab_libfunc (ne_optab, SFmode, NULL); ++ set_optab_libfunc (lt_optab, SFmode, "__avr32_f32_cmp_lt"); ++ set_optab_libfunc (le_optab, SFmode, NULL); ++ set_optab_libfunc (ge_optab, SFmode, "__avr32_f32_cmp_ge"); ++ set_optab_libfunc (gt_optab, SFmode, NULL); ++ ++ /* Floating-point to integer conversions. */ ++ set_conv_libfunc (sfix_optab, SImode, DFmode, "__avr32_f64_to_s32"); ++ set_conv_libfunc (ufix_optab, SImode, DFmode, "__avr32_f64_to_u32"); ++ set_conv_libfunc (sfix_optab, DImode, DFmode, "__avr32_f64_to_s64"); ++ set_conv_libfunc (ufix_optab, DImode, DFmode, "__avr32_f64_to_u64"); ++ set_conv_libfunc (sfix_optab, SImode, SFmode, "__avr32_f32_to_s32"); ++ set_conv_libfunc (ufix_optab, SImode, SFmode, "__avr32_f32_to_u32"); ++ set_conv_libfunc (sfix_optab, DImode, SFmode, "__avr32_f32_to_s64"); ++ set_conv_libfunc (ufix_optab, DImode, SFmode, "__avr32_f32_to_u64"); ++ ++ /* Conversions between floating types. */ ++ set_conv_libfunc (trunc_optab, SFmode, DFmode, "__avr32_f64_to_f32"); ++ set_conv_libfunc (sext_optab, DFmode, SFmode, "__avr32_f32_to_f64"); ++ ++ /* Integer to floating-point conversions. Table 8. */ ++ set_conv_libfunc (sfloat_optab, DFmode, SImode, "__avr32_s32_to_f64"); ++ set_conv_libfunc (sfloat_optab, DFmode, DImode, "__avr32_s64_to_f64"); ++ set_conv_libfunc (sfloat_optab, SFmode, SImode, "__avr32_s32_to_f32"); ++ set_conv_libfunc (sfloat_optab, SFmode, DImode, "__avr32_s64_to_f32"); ++ set_conv_libfunc (ufloat_optab, DFmode, SImode, "__avr32_u32_to_f64"); ++ set_conv_libfunc (ufloat_optab, SFmode, SImode, "__avr32_u32_to_f32"); ++ /* TODO: Add these to gcc library functions */ ++ //set_conv_libfunc (ufloat_optab, DFmode, DImode, NULL); ++ //set_conv_libfunc (ufloat_optab, SFmode, DImode, NULL); ++ ++ /* Long long. Table 9. */ ++ set_optab_libfunc (smul_optab, DImode, "__avr32_mul64"); ++ set_optab_libfunc (sdiv_optab, DImode, "__avr32_sdiv64"); ++ set_optab_libfunc (udiv_optab, DImode, "__avr32_udiv64"); ++ set_optab_libfunc (smod_optab, DImode, "__avr32_smod64"); ++ set_optab_libfunc (umod_optab, DImode, "__avr32_umod64"); ++ set_optab_libfunc (ashl_optab, DImode, "__avr32_lsl64"); ++ set_optab_libfunc (lshr_optab, DImode, "__avr32_lsr64"); ++ set_optab_libfunc (ashr_optab, DImode, "__avr32_asr64"); ++ ++ /* Floating point library functions which have fast versions. */ ++ if ( TARGET_FAST_FLOAT ) ++ { ++ set_optab_libfunc (add_optab, DFmode, "__avr32_f64_add_fast"); ++ set_optab_libfunc (sub_optab, DFmode, "__avr32_f64_sub_fast"); ++ set_optab_libfunc (add_optab, SFmode, "__avr32_f32_add_fast"); ++ set_optab_libfunc (sub_optab, SFmode, "__avr32_f32_sub_fast"); ++ set_optab_libfunc (sdiv_optab, SFmode, "__avr32_f32_div_fast"); ++ } ++ else ++ { ++ set_optab_libfunc (add_optab, DFmode, "__avr32_f64_add"); ++ set_optab_libfunc (sub_optab, DFmode, "__avr32_f64_sub"); ++ set_optab_libfunc (add_optab, SFmode, "__avr32_f32_add"); ++ set_optab_libfunc (sub_optab, SFmode, "__avr32_f32_sub"); ++ set_optab_libfunc (sdiv_optab, SFmode, "__avr32_f32_div"); ++ } ++} ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-elf.h +--- gcc-4.2.1/gcc/config/avr32/avr32-elf.h 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-elf.h 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,84 @@ ++/* ++ Elf specific definitions. ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++ ++/***************************************************************************** ++ * Controlling the Compilator Driver, 'gcc' ++ *****************************************************************************/ ++ ++/* Run-time Target Specification. */ ++#undef TARGET_VERSION ++#define TARGET_VERSION fputs (" (AVR32 GNU with ELF)", stderr); ++ ++/* ++Another C string constant used much like LINK_SPEC. The ++difference between the two is that STARTFILE_SPEC is used at ++the very beginning of the command given to the linker. ++ ++If this macro is not defined, a default is provided that loads the ++standard C startup file from the usual place. See gcc.c. ++*/ ++#undef STARTFILE_SPEC ++#define STARTFILE_SPEC "crt0%O%s crti%O%s crtbegin%O%s" ++ ++#undef LINK_SPEC ++#define LINK_SPEC "%{muse-oscall:--defsym __do_not_use_oscall_coproc__=0} %{mrelax|O*:%{mno-relax|O0|O1: ;:--relax}} %{mpart=*:-mavr32elf_%*} %{mcpu=*:-mavr32elf_%*}" ++ ++ ++/* ++Another C string constant used much like LINK_SPEC. The ++difference between the two is that ENDFILE_SPEC is used at ++the very end of the command given to the linker. ++ ++Do not define this macro if it does not need to do anything. ++*/ ++#undef ENDFILE_SPEC ++#define ENDFILE_SPEC "crtend%O%s crtn%O%s" ++ ++ ++/* Target CPU builtins. */ ++#define TARGET_CPU_CPP_BUILTINS() \ ++ do \ ++ { \ ++ builtin_define ("__avr32__"); \ ++ builtin_define ("__AVR32__"); \ ++ builtin_define ("__AVR32_ELF__"); \ ++ builtin_define (avr32_part->macro); \ ++ builtin_define (avr32_arch->macro); \ ++ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) \ ++ builtin_define ("__AVR32_AVR32A__"); \ ++ else \ ++ builtin_define ("__AVR32_AVR32B__"); \ ++ if (TARGET_UNALIGNED_WORD) \ ++ builtin_define ("__AVR32_HAS_UNALIGNED_WORD__"); \ ++ if (TARGET_SIMD) \ ++ builtin_define ("__AVR32_HAS_SIMD__"); \ ++ if (TARGET_DSP) \ ++ builtin_define ("__AVR32_HAS_DSP__"); \ ++ if (TARGET_RMW) \ ++ builtin_define ("__AVR32_HAS_RMW__"); \ ++ if (TARGET_BRANCH_PRED) \ ++ builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ ++ if (TARGET_FAST_FLOAT) \ ++ builtin_define ("__AVR32_FAST_FLOAT__"); \ ++ } \ ++ while (0) +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.h +--- gcc-4.2.1/gcc/config/avr32/avr32.h 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.h 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,3281 @@ ++/* ++ Definitions of target machine for AVR32. ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ Initial porting by Anders �dland. ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++#ifndef GCC_AVR32_H ++#define GCC_AVR32_H ++ ++ ++#ifndef OBJECT_FORMAT_ELF ++#error avr32.h included before elfos.h ++#endif ++ ++#ifndef LOCAL_LABEL_PREFIX ++#define LOCAL_LABEL_PREFIX "." ++#endif ++ ++#ifndef SUBTARGET_CPP_SPEC ++#define SUBTARGET_CPP_SPEC "-D__ELF__" ++#endif ++ ++ ++extern struct rtx_def *avr32_compare_op0; ++extern struct rtx_def *avr32_compare_op1; ++ ++ ++extern struct rtx_def *avr32_acc_cache; ++ ++/* cache instruction op5 codes */ ++#define AVR32_CACHE_INVALIDATE_ICACHE 1 ++ ++/* These bits describe the different types of function supported ++ by the AVR32 backend. They are exclusive. ie a function cannot be both a ++ normal function and an interworked function, for example. Knowing the ++ type of a function is important for determining its prologue and ++ epilogue sequences. ++ Note value 7 is currently unassigned. Also note that the interrupt ++ function types all have bit 2 set, so that they can be tested for easily. ++ Note that 0 is deliberately chosen for AVR32_FT_UNKNOWN so that when the ++ machine_function structure is initialized (to zero) func_type will ++ default to unknown. This will force the first use of avr32_current_func_type ++ to call avr32_compute_func_type. */ ++#define AVR32_FT_UNKNOWN 0 /* Type has not yet been determined. ++ */ ++#define AVR32_FT_NORMAL 1 /* Your normal, straightforward ++ function. */ ++#define AVR32_FT_ACALL 2 /* An acall function. */ ++#define AVR32_FT_EXCEPTION_HANDLER 3 /* A C++ exception handler. */ ++#define AVR32_FT_ISR_FULL 4 /* A fully shadowed interrupt mode. */ ++#define AVR32_FT_ISR_HALF 5 /* A half shadowed interrupt mode. */ ++#define AVR32_FT_ISR_NONE 6 /* No shadow registers. */ ++ ++#define AVR32_FT_TYPE_MASK ((1 << 3) - 1) ++ ++/* In addition functions can have several type modifiers, ++ outlined by these bit masks: */ ++#define AVR32_FT_INTERRUPT (1 << 2) /* Note overlap with FT_ISR ++ and above. */ ++#define AVR32_FT_NAKED (1 << 3) /* No prologue or epilogue. */ ++#define AVR32_FT_VOLATILE (1 << 4) /* Does not return. */ ++#define AVR32_FT_NESTED (1 << 5) /* Embedded inside another ++ func. */ ++ ++/* Some macros to test these flags. */ ++#define AVR32_FUNC_TYPE(t) (t & AVR32_FT_TYPE_MASK) ++#define IS_INTERRUPT(t) (t & AVR32_FT_INTERRUPT) ++#define IS_VOLATILE(t) (t & AVR32_FT_VOLATILE) ++#define IS_NAKED(t) (t & AVR32_FT_NAKED) ++#define IS_NESTED(t) (t & AVR32_FT_NESTED) ++ ++ ++typedef struct minipool_labels ++GTY ((chain_next ("%h.next"), chain_prev ("%h.prev"))) ++{ ++ rtx label; ++ struct minipool_labels *prev; ++ struct minipool_labels *next; ++} minipool_labels; ++ ++/* A C structure for machine-specific, per-function data. ++ This is added to the cfun structure. */ ++ ++typedef struct machine_function ++GTY (()) ++{ ++ /* Records the type of the current function. */ ++ unsigned long func_type; ++ /* List of minipool labels, use for checking if code label is valid in a ++ memory expression */ ++ minipool_labels *minipool_label_head; ++ minipool_labels *minipool_label_tail; ++} machine_function; ++ ++/* Initialize data used by insn expanders. This is called from insn_emit, ++ once for every function before code is generated. */ ++#define INIT_EXPANDERS avr32_init_expanders () ++ ++/****************************************************************************** ++ * SPECS ++ *****************************************************************************/ ++ ++#ifndef ASM_SPEC ++#define ASM_SPEC "%{fpic:--pic} %{mrelax|O*:%{mno-relax|O0|O1: ;:--linkrelax}} %{march=*:-march=%*} %{mpart=*:-mpart=%*}" ++#endif ++ ++#ifndef MULTILIB_DEFAULTS ++#define MULTILIB_DEFAULTS { "march=ap", "" } ++#endif ++ ++/****************************************************************************** ++ * Run-time Target Specification ++ *****************************************************************************/ ++#ifndef TARGET_VERSION ++#define TARGET_VERSION fprintf(stderr, " (AVR32, GNU assembler syntax)"); ++#endif ++ ++/* Part types. Keep this in sync with the order of avr32_part_types in avr32.c*/ ++enum part_type ++{ ++ PART_TYPE_AVR32_NONE, ++ PART_TYPE_AVR32_AP7000, ++ PART_TYPE_AVR32_AP7010, ++ PART_TYPE_AVR32_AP7020, ++ PART_TYPE_AVR32_UC3A0256, ++ PART_TYPE_AVR32_UC3A0512, ++ PART_TYPE_AVR32_UC3A1128, ++ PART_TYPE_AVR32_UC3A1256, ++ PART_TYPE_AVR32_UC3A1512, ++ PART_TYPE_AVR32_UC3B064, ++ PART_TYPE_AVR32_UC3B0128, ++ PART_TYPE_AVR32_UC3B0256, ++ PART_TYPE_AVR32_UC3B164, ++ PART_TYPE_AVR32_UC3B1128, ++ PART_TYPE_AVR32_UC3B1256 ++}; ++ ++/* Microarchitectures. */ ++enum microarchitecture_type ++{ ++ UARCH_TYPE_AVR32A, ++ UARCH_TYPE_AVR32B, ++ UARCH_TYPE_NONE ++}; ++ ++/* Architectures types which specifies the pipeline. ++ Keep this in sync with avr32_arch_types in avr32.c ++ and the pipeline attribute in avr32.md */ ++enum architecture_type ++{ ++ ARCH_TYPE_AVR32_AP, ++ ARCH_TYPE_AVR32_UC, ++ ARCH_TYPE_AVR32_NONE ++}; ++ ++/* Flag specifying if the cpu has support for DSP instructions.*/ ++#define FLAG_AVR32_HAS_DSP (1 << 0) ++/* Flag specifying if the cpu has support for Read-Modify-Write ++ instructions.*/ ++#define FLAG_AVR32_HAS_RMW (1 << 1) ++/* Flag specifying if the cpu has support for SIMD instructions. */ ++#define FLAG_AVR32_HAS_SIMD (1 << 2) ++/* Flag specifying if the cpu has support for unaligned memory word access. */ ++#define FLAG_AVR32_HAS_UNALIGNED_WORD (1 << 3) ++/* Flag specifying if the cpu has support for branch prediction. */ ++#define FLAG_AVR32_HAS_BRANCH_PRED (1 << 4) ++/* Flag specifying if the cpu has support for a return stack. */ ++#define FLAG_AVR32_HAS_RETURN_STACK (1 << 5) ++/* Flag specifying if the cpu has caches. */ ++#define FLAG_AVR32_HAS_CACHES (1 << 6) ++ ++/* Structure for holding information about different avr32 CPUs/parts */ ++struct part_type_s ++{ ++ const char *const name; ++ enum part_type part_type; ++ enum architecture_type arch_type; ++ /* Must lie outside user's namespace. NULL == no macro. */ ++ const char *const macro; ++}; ++ ++/* Structure for holding information about different avr32 pipeline ++ architectures. */ ++struct arch_type_s ++{ ++ const char *const name; ++ enum architecture_type arch_type; ++ enum microarchitecture_type uarch_type; ++ const unsigned long feature_flags; ++ /* Must lie outside user's namespace. NULL == no macro. */ ++ const char *const macro; ++}; ++ ++extern const struct part_type_s *avr32_part; ++extern const struct arch_type_s *avr32_arch; ++ ++#define TARGET_SIMD (avr32_arch->feature_flags & FLAG_AVR32_HAS_SIMD) ++#define TARGET_DSP (avr32_arch->feature_flags & FLAG_AVR32_HAS_DSP) ++#define TARGET_RMW (avr32_arch->feature_flags & FLAG_AVR32_HAS_RMW) ++#define TARGET_UNALIGNED_WORD (avr32_arch->feature_flags & FLAG_AVR32_HAS_UNALIGNED_WORD) ++#define TARGET_BRANCH_PRED (avr32_arch->feature_flags & FLAG_AVR32_HAS_BRANCH_PRED) ++#define TARGET_RETURN_STACK (avr32_arch->feature_flags & FLAG_AVR32_HAS_RETURN_STACK) ++#define TARGET_CACHES (avr32_arch->feature_flags & FLAG_AVR32_HAS_CACHES) ++ ++#define CAN_DEBUG_WITHOUT_FP ++ ++/****************************************************************************** ++ * Storage Layout ++ *****************************************************************************/ ++ ++/* ++Define this macro to have the value 1 if the most significant bit in a ++byte has the lowest number; otherwise define it to have the value zero. ++This means that bit-field instructions count from the most significant ++bit. If the machine has no bit-field instructions, then this must still ++be defined, but it doesn't matter which value it is defined to. This ++macro need not be a constant. ++ ++This macro does not affect the way structure fields are packed into ++bytes or words; that is controlled by BYTES_BIG_ENDIAN. ++*/ ++#define BITS_BIG_ENDIAN 0 ++ ++/* ++Define this macro to have the value 1 if the most significant byte in a ++word has the lowest number. This macro need not be a constant. ++*/ ++/* ++ Data is stored in an big-endian way. ++*/ ++#define BYTES_BIG_ENDIAN 1 ++ ++/* ++Define this macro to have the value 1 if, in a multiword object, the ++most significant word has the lowest number. This applies to both ++memory locations and registers; GCC fundamentally assumes that the ++order of words in memory is the same as the order in registers. This ++macro need not be a constant. ++*/ ++/* ++ Data is stored in an bin-endian way. ++*/ ++#define WORDS_BIG_ENDIAN 1 ++ ++/* ++Define this macro if WORDS_BIG_ENDIAN is not constant. This must be a ++constant value with the same meaning as WORDS_BIG_ENDIAN, which will be ++used only when compiling libgcc2.c. Typically the value will be set ++based on preprocessor defines. ++*/ ++#define LIBGCC2_WORDS_BIG_ENDIAN WORDS_BIG_ENDIAN ++ ++/* ++Define this macro to have the value 1 if DFmode, XFmode or ++TFmode floating point numbers are stored in memory with the word ++containing the sign bit at the lowest address; otherwise define it to ++have the value 0. This macro need not be a constant. ++ ++You need not define this macro if the ordering is the same as for ++multi-word integers. ++*/ ++/* #define FLOAT_WORDS_BIG_ENDIAN 1 */ ++ ++/* ++Define this macro to be the number of bits in an addressable storage ++unit (byte); normally 8. ++*/ ++#define BITS_PER_UNIT 8 ++ ++/* ++Number of bits in a word; normally 32. ++*/ ++#define BITS_PER_WORD 32 ++ ++/* ++Maximum number of bits in a word. If this is undefined, the default is ++BITS_PER_WORD. Otherwise, it is the constant value that is the ++largest value that BITS_PER_WORD can have at run-time. ++*/ ++/* MAX_BITS_PER_WORD not defined*/ ++ ++/* ++Number of storage units in a word; normally 4. ++*/ ++#define UNITS_PER_WORD 4 ++ ++/* ++Minimum number of units in a word. If this is undefined, the default is ++UNITS_PER_WORD. Otherwise, it is the constant value that is the ++smallest value that UNITS_PER_WORD can have at run-time. ++*/ ++/* MIN_UNITS_PER_WORD not defined */ ++ ++/* ++Width of a pointer, in bits. You must specify a value no wider than the ++width of Pmode. If it is not equal to the width of Pmode, ++you must define POINTERS_EXTEND_UNSIGNED. ++*/ ++#define POINTER_SIZE 32 ++ ++/* ++A C expression whose value is greater than zero if pointers that need to be ++extended from being POINTER_SIZE bits wide to Pmode are to ++be zero-extended and zero if they are to be sign-extended. If the value ++is less then zero then there must be an "ptr_extend" instruction that ++extends a pointer from POINTER_SIZE to Pmode. ++ ++You need not define this macro if the POINTER_SIZE is equal ++to the width of Pmode. ++*/ ++/* #define POINTERS_EXTEND_UNSIGNED */ ++ ++/* ++A Macro to update M and UNSIGNEDP when an object whose type ++is TYPE and which has the specified mode and signedness is to be ++stored in a register. This macro is only called when TYPE is a ++scalar type. ++ ++On most RISC machines, which only have operations that operate on a full ++register, define this macro to set M to word_mode if ++M is an integer mode narrower than BITS_PER_WORD. In most ++cases, only integer modes should be widened because wider-precision ++floating-point operations are usually more expensive than their narrower ++counterparts. ++ ++For most machines, the macro definition does not change UNSIGNEDP. ++However, some machines, have instructions that preferentially handle ++either signed or unsigned quantities of certain modes. For example, on ++the DEC Alpha, 32-bit loads from memory and 32-bit add instructions ++sign-extend the result to 64 bits. On such machines, set ++UNSIGNEDP according to which kind of extension is more efficient. ++ ++Do not define this macro if it would never modify M. ++*/ ++#define PROMOTE_MODE(M, UNSIGNEDP, TYPE) \ ++ { \ ++ if (GET_MODE_CLASS (M) == MODE_INT \ ++ && GET_MODE_SIZE (M) < 4) \ ++ { \ ++ if (M == QImode) \ ++ UNSIGNEDP = 1; \ ++ else if (M == SImode) \ ++ UNSIGNEDP = 0; \ ++ (M) = SImode; \ ++ } \ ++ } ++ ++#define PROMOTE_FUNCTION_MODE(M, UNSIGNEDP, TYPE) \ ++ { \ ++ if (GET_MODE_CLASS (M) == MODE_INT \ ++ && GET_MODE_SIZE (M) < 4) \ ++ { \ ++ (M) = SImode; \ ++ } \ ++ } ++ ++/* Define if operations between registers always perform the operation ++ on the full register even if a narrower mode is specified. */ ++#define WORD_REGISTER_OPERATIONS ++ ++/* Define if loading in MODE, an integral mode narrower than BITS_PER_WORD ++ will either zero-extend or sign-extend. The value of this macro should ++ be the code that says which one of the two operations is implicitly ++ done, UNKNOWN if not known. */ ++#define LOAD_EXTEND_OP(MODE) \ ++ (((MODE) == QImode) ? ZERO_EXTEND \ ++ : ((MODE) == HImode) ? SIGN_EXTEND : UNKNOWN) ++ ++ ++/* ++Define this macro if the promotion described by PROMOTE_MODE ++should only be performed for outgoing function arguments or ++function return values, as specified by PROMOTE_FUNCTION_ARGS ++and PROMOTE_FUNCTION_RETURN, respectively. ++*/ ++/* #define PROMOTE_FOR_CALL_ONLY */ ++ ++/* ++Normal alignment required for function parameters on the stack, in ++bits. All stack parameters receive at least this much alignment ++regardless of data type. On most machines, this is the same as the ++size of an integer. ++*/ ++#define PARM_BOUNDARY 32 ++ ++/* ++Define this macro to the minimum alignment enforced by hardware for the ++stack pointer on this machine. The definition is a C expression for the ++desired alignment (measured in bits). This value is used as a default ++if PREFERRED_STACK_BOUNDARY is not defined. On most machines, ++this should be the same as PARM_BOUNDARY. ++*/ ++#define STACK_BOUNDARY 32 ++ ++/* ++Define this macro if you wish to preserve a certain alignment for the ++stack pointer, greater than what the hardware enforces. The definition ++is a C expression for the desired alignment (measured in bits). This ++macro must evaluate to a value equal to or larger than ++STACK_BOUNDARY. ++*/ ++#define PREFERRED_STACK_BOUNDARY (TARGET_FORCE_DOUBLE_ALIGN ? 64 : 32 ) ++ ++/* ++Alignment required for a function entry point, in bits. ++*/ ++#define FUNCTION_BOUNDARY 16 ++ ++/* ++Biggest alignment that any data type can require on this machine, in bits. ++*/ ++#define BIGGEST_ALIGNMENT (TARGET_FORCE_DOUBLE_ALIGN ? 64 : 32 ) ++ ++/* ++If defined, the smallest alignment, in bits, that can be given to an ++object that can be referenced in one operation, without disturbing any ++nearby object. Normally, this is BITS_PER_UNIT, but may be larger ++on machines that don't have byte or half-word store operations. ++*/ ++#define MINIMUM_ATOMIC_ALIGNMENT BITS_PER_UNIT ++ ++ ++/* ++An integer expression for the size in bits of the largest integer machine mode that ++should actually be used. All integer machine modes of this size or smaller can be ++used for structures and unions with the appropriate sizes. If this macro is undefined, ++GET_MODE_BITSIZE (DImode) is assumed.*/ ++#define MAX_FIXED_MODE_SIZE GET_MODE_BITSIZE (DImode) ++ ++ ++/* ++If defined, a C expression to compute the alignment given to a constant ++that is being placed in memory. CONSTANT is the constant and ++BASIC_ALIGN is the alignment that the object would ordinarily ++have. The value of this macro is used instead of that alignment to ++align the object. ++ ++If this macro is not defined, then BASIC_ALIGN is used. ++ ++The typical use of this macro is to increase alignment for string ++constants to be word aligned so that strcpy calls that copy ++constants can be done inline. ++*/ ++#define CONSTANT_ALIGNMENT(CONSTANT, BASIC_ALIGN) \ ++ ((TREE_CODE(CONSTANT) == STRING_CST) ? BITS_PER_WORD : BASIC_ALIGN) ++ ++/* Try to align string to a word. */ ++#define DATA_ALIGNMENT(TYPE, ALIGN) \ ++ ({(TREE_CODE (TYPE) == ARRAY_TYPE \ ++ && TYPE_MODE (TREE_TYPE (TYPE)) == QImode \ ++ && (ALIGN) < BITS_PER_WORD ? BITS_PER_WORD : (ALIGN));}) ++ ++/* Try to align local store strings to a word. */ ++#define LOCAL_ALIGNMENT(TYPE, ALIGN) \ ++ ({(TREE_CODE (TYPE) == ARRAY_TYPE \ ++ && TYPE_MODE (TREE_TYPE (TYPE)) == QImode \ ++ && (ALIGN) < BITS_PER_WORD ? BITS_PER_WORD : (ALIGN));}) ++ ++/* ++Define this macro to be the value 1 if instructions will fail to work ++if given data not on the nominal alignment. If instructions will merely ++go slower in that case, define this macro as 0. ++*/ ++#define STRICT_ALIGNMENT 1 ++ ++/* ++Define this if you wish to imitate the way many other C compilers handle ++alignment of bit-fields and the structures that contain them. ++ ++The behavior is that the type written for a bit-field (int, ++short, or other integer type) imposes an alignment for the ++entire structure, as if the structure really did contain an ordinary ++field of that type. In addition, the bit-field is placed within the ++structure so that it would fit within such a field, not crossing a ++boundary for it. ++ ++Thus, on most machines, a bit-field whose type is written as int ++would not cross a four-byte boundary, and would force four-byte ++alignment for the whole structure. (The alignment used may not be four ++bytes; it is controlled by the other alignment parameters.) ++ ++If the macro is defined, its definition should be a C expression; ++a nonzero value for the expression enables this behavior. ++ ++Note that if this macro is not defined, or its value is zero, some ++bit-fields may cross more than one alignment boundary. The compiler can ++support such references if there are insv, extv, and ++extzv insns that can directly reference memory. ++ ++The other known way of making bit-fields work is to define ++STRUCTURE_SIZE_BOUNDARY as large as BIGGEST_ALIGNMENT. ++Then every structure can be accessed with fullwords. ++ ++Unless the machine has bit-field instructions or you define ++STRUCTURE_SIZE_BOUNDARY that way, you must define ++PCC_BITFIELD_TYPE_MATTERS to have a nonzero value. ++ ++If your aim is to make GCC use the same conventions for laying out ++bit-fields as are used by another compiler, here is how to investigate ++what the other compiler does. Compile and run this program: ++ ++struct foo1 ++{ ++ char x; ++ char :0; ++ char y; ++}; ++ ++struct foo2 ++{ ++ char x; ++ int :0; ++ char y; ++}; ++ ++main () ++{ ++ printf ("Size of foo1 is %d\n", ++ sizeof (struct foo1)); ++ printf ("Size of foo2 is %d\n", ++ sizeof (struct foo2)); ++ exit (0); ++} ++ ++If this prints 2 and 5, then the compiler's behavior is what you would ++get from PCC_BITFIELD_TYPE_MATTERS. ++*/ ++#define PCC_BITFIELD_TYPE_MATTERS 1 ++ ++ ++/****************************************************************************** ++ * Layout of Source Language Data Types ++ *****************************************************************************/ ++ ++/* ++A C expression for the size in bits of the type int on the ++target machine. If you don't define this, the default is one word. ++*/ ++#define INT_TYPE_SIZE 32 ++ ++/* ++A C expression for the size in bits of the type short on the ++target machine. If you don't define this, the default is half a word. (If ++this would be less than one storage unit, it is rounded up to one unit.) ++*/ ++#define SHORT_TYPE_SIZE 16 ++ ++/* ++A C expression for the size in bits of the type long on the ++target machine. If you don't define this, the default is one word. ++*/ ++#define LONG_TYPE_SIZE 32 ++ ++ ++/* ++A C expression for the size in bits of the type long long on the ++target machine. If you don't define this, the default is two ++words. If you want to support GNU Ada on your machine, the value of this ++macro must be at least 64. ++*/ ++#define LONG_LONG_TYPE_SIZE 64 ++ ++/* ++A C expression for the size in bits of the type char on the ++target machine. If you don't define this, the default is ++BITS_PER_UNIT. ++*/ ++#define CHAR_TYPE_SIZE 8 ++ ++ ++/* ++A C expression for the size in bits of the C++ type bool and ++C99 type _Bool on the target machine. If you don't define ++this, and you probably shouldn't, the default is CHAR_TYPE_SIZE. ++*/ ++#define BOOL_TYPE_SIZE 8 ++ ++ ++/* ++An expression whose value is 1 or 0, according to whether the type ++char should be signed or unsigned by default. The user can ++always override this default with the options -fsigned-char ++and -funsigned-char. ++*/ ++/* We are using unsigned char */ ++#define DEFAULT_SIGNED_CHAR 0 ++ ++ ++/* ++A C expression for a string describing the name of the data type to use ++for size values. The typedef name size_t is defined using the ++contents of the string. ++ ++The string can contain more than one keyword. If so, separate them with ++spaces, and write first any length keyword, then unsigned if ++appropriate, and finally int. The string must exactly match one ++of the data type names defined in the function ++init_decl_processing in the file c-decl.c. You may not ++omit int or change the order - that would cause the compiler to ++crash on startup. ++ ++If you don't define this macro, the default is "long unsigned int". ++*/ ++#define SIZE_TYPE "long unsigned int" ++ ++/* ++A C expression for a string describing the name of the data type to use ++for the result of subtracting two pointers. The typedef name ++ptrdiff_t is defined using the contents of the string. See ++SIZE_TYPE above for more information. ++ ++If you don't define this macro, the default is "long int". ++*/ ++#define PTRDIFF_TYPE "long int" ++ ++ ++/* ++A C expression for the size in bits of the data type for wide ++characters. This is used in cpp, which cannot make use of ++WCHAR_TYPE. ++*/ ++#define WCHAR_TYPE_SIZE 32 ++ ++ ++/* ++A C expression for a string describing the name of the data type to ++use for wide characters passed to printf and returned from ++getwc. The typedef name wint_t is defined using the ++contents of the string. See SIZE_TYPE above for more ++information. ++ ++If you don't define this macro, the default is "unsigned int". ++*/ ++#define WINT_TYPE "unsigned int" ++ ++/* ++A C expression for a string describing the name of the data type that ++can represent any value of any standard or extended signed integer type. ++The typedef name intmax_t is defined using the contents of the ++string. See SIZE_TYPE above for more information. ++ ++If you don't define this macro, the default is the first of ++"int", "long int", or "long long int" that has as ++much precision as long long int. ++*/ ++#define INTMAX_TYPE "long long int" ++ ++/* ++A C expression for a string describing the name of the data type that ++can represent any value of any standard or extended unsigned integer ++type. The typedef name uintmax_t is defined using the contents ++of the string. See SIZE_TYPE above for more information. ++ ++If you don't define this macro, the default is the first of ++"unsigned int", "long unsigned int", or "long long unsigned int" ++that has as much precision as long long unsigned int. ++*/ ++#define UINTMAX_TYPE "long long unsigned int" ++ ++ ++/****************************************************************************** ++ * Register Usage ++ *****************************************************************************/ ++ ++/* Convert from gcc internal register number to register number ++ used in assembly code */ ++#define ASM_REGNUM(reg) (LAST_REGNUM - (reg)) ++#define ASM_FP_REGNUM(reg) (LAST_FP_REGNUM - (reg)) ++ ++/* Convert between register number used in assembly to gcc ++ internal register number */ ++#define INTERNAL_REGNUM(reg) (LAST_REGNUM - (reg)) ++#define INTERNAL_FP_REGNUM(reg) (LAST_FP_REGNUM - (reg)) ++ ++/** Basic Characteristics of Registers **/ ++ ++/* ++Number of hardware registers known to the compiler. They receive ++numbers 0 through FIRST_PSEUDO_REGISTER-1; thus, the first ++pseudo register's number really is assigned the number ++FIRST_PSEUDO_REGISTER. ++*/ ++#define FIRST_PSEUDO_REGISTER (LAST_FP_REGNUM + 1) ++ ++#define FIRST_REGNUM 0 ++#define LAST_REGNUM 15 ++#define NUM_FP_REGS 16 ++#define FIRST_FP_REGNUM 16 ++#define LAST_FP_REGNUM (16+NUM_FP_REGS-1) ++ ++/* ++An initializer that says which registers are used for fixed purposes ++all throughout the compiled code and are therefore not available for ++general allocation. These would include the stack pointer, the frame ++pointer (except on machines where that can be used as a general ++register when no frame pointer is needed), the program counter on ++machines where that is considered one of the addressable registers, ++and any other numbered register with a standard use. ++ ++This information is expressed as a sequence of numbers, separated by ++commas and surrounded by braces. The nth number is 1 if ++register n is fixed, 0 otherwise. ++ ++The table initialized from this macro, and the table initialized by ++the following one, may be overridden at run time either automatically, ++by the actions of the macro CONDITIONAL_REGISTER_USAGE, or by ++the user with the command options -ffixed-[reg], ++-fcall-used-[reg] and -fcall-saved-[reg]. ++*/ ++ ++/* The internal gcc register numbers are reversed ++ compared to the real register numbers since ++ gcc expects data types stored over multiple ++ registers in the register file to be big endian ++ if the memory layout is big endian. But this ++ is not the case for avr32 so we fake a big ++ endian register file. */ ++ ++#define FIXED_REGISTERS { \ ++ 1, /* Program Counter */ \ ++ 0, /* Link Register */ \ ++ 1, /* Stack Pointer */ \ ++ 0, /* r12 */ \ ++ 0, /* r11 */ \ ++ 0, /* r10 */ \ ++ 0, /* r9 */ \ ++ 0, /* r8 */ \ ++ 0, /* r7 */ \ ++ 0, /* r6 */ \ ++ 0, /* r5 */ \ ++ 0, /* r4 */ \ ++ 0, /* r3 */ \ ++ 0, /* r2 */ \ ++ 0, /* r1 */ \ ++ 0, /* r0 */ \ ++ 0, /* f15 */ \ ++ 0, /* f14 */ \ ++ 0, /* f13 */ \ ++ 0, /* f12 */ \ ++ 0, /* f11 */ \ ++ 0, /* f10 */ \ ++ 0, /* f9 */ \ ++ 0, /* f8 */ \ ++ 0, /* f7 */ \ ++ 0, /* f6 */ \ ++ 0, /* f5 */ \ ++ 0, /* f4 */ \ ++ 0, /* f3 */ \ ++ 0, /* f2*/ \ ++ 0, /* f1 */ \ ++ 0 /* f0 */ \ ++} ++ ++/* ++Like FIXED_REGISTERS but has 1 for each register that is ++clobbered (in general) by function calls as well as for fixed ++registers. This macro therefore identifies the registers that are not ++available for general allocation of values that must live across ++function calls. ++ ++If a register has 0 in CALL_USED_REGISTERS, the compiler ++automatically saves it on function entry and restores it on function ++exit, if the register is used within the function. ++*/ ++#define CALL_USED_REGISTERS { \ ++ 1, /* Program Counter */ \ ++ 0, /* Link Register */ \ ++ 1, /* Stack Pointer */ \ ++ 1, /* r12 */ \ ++ 1, /* r11 */ \ ++ 1, /* r10 */ \ ++ 1, /* r9 */ \ ++ 1, /* r8 */ \ ++ 0, /* r7 */ \ ++ 0, /* r6 */ \ ++ 0, /* r5 */ \ ++ 0, /* r4 */ \ ++ 0, /* r3 */ \ ++ 0, /* r2 */ \ ++ 0, /* r1 */ \ ++ 0, /* r0 */ \ ++ 1, /* f15 */ \ ++ 1, /* f14 */ \ ++ 1, /* f13 */ \ ++ 1, /* f12 */ \ ++ 1, /* f11 */ \ ++ 1, /* f10 */ \ ++ 1, /* f9 */ \ ++ 1, /* f8 */ \ ++ 0, /* f7 */ \ ++ 0, /* f6 */ \ ++ 0, /* f5 */ \ ++ 0, /* f4 */ \ ++ 0, /* f3 */ \ ++ 0, /* f2*/ \ ++ 0, /* f1*/ \ ++ 0, /* f0 */ \ ++} ++ ++/* Interrupt functions can only use registers that have already been ++ saved by the prologue, even if they would normally be ++ call-clobbered. */ ++#define HARD_REGNO_RENAME_OK(SRC, DST) \ ++ (! IS_INTERRUPT (cfun->machine->func_type) || \ ++ regs_ever_live[DST]) ++ ++ ++/* ++Zero or more C statements that may conditionally modify five variables ++fixed_regs, call_used_regs, global_regs, ++reg_names, and reg_class_contents, to take into account ++any dependence of these register sets on target flags. The first three ++of these are of type char [] (interpreted as Boolean vectors). ++global_regs is a const char *[], and ++reg_class_contents is a HARD_REG_SET. Before the macro is ++called, fixed_regs, call_used_regs, ++reg_class_contents, and reg_names have been initialized ++from FIXED_REGISTERS, CALL_USED_REGISTERS, ++REG_CLASS_CONTENTS, and REGISTER_NAMES, respectively. ++global_regs has been cleared, and any -ffixed-[reg], ++-fcall-used-[reg] and -fcall-saved-[reg] ++command options have been applied. ++ ++You need not define this macro if it has no work to do. ++ ++If the usage of an entire class of registers depends on the target ++flags, you may indicate this to GCC by using this macro to modify ++fixed_regs and call_used_regs to 1 for each of the ++registers in the classes which should not be used by GCC. Also define ++the macro REG_CLASS_FROM_LETTER to return NO_REGS if it ++is called with a letter for a class that shouldn't be used. ++ ++ (However, if this class is not included in GENERAL_REGS and all ++of the insn patterns whose constraints permit this class are ++controlled by target switches, then GCC will automatically avoid using ++these registers when the target switches are opposed to them.) ++*/ ++#define CONDITIONAL_REGISTER_USAGE \ ++ do \ ++ { \ ++ int regno; \ ++ \ ++ if (TARGET_SOFT_FLOAT) \ ++ { \ ++ for (regno = FIRST_FP_REGNUM; \ ++ regno <= LAST_FP_REGNUM; ++regno) \ ++ fixed_regs[regno] = call_used_regs[regno] = 1; \ ++ } \ ++ if (flag_pic) \ ++ { \ ++ fixed_regs[PIC_OFFSET_TABLE_REGNUM] = 1; \ ++ call_used_regs[PIC_OFFSET_TABLE_REGNUM] = 1; \ ++ } \ ++ } \ ++ while (0) ++ ++ ++/* ++If the program counter has a register number, define this as that ++register number. Otherwise, do not define it. ++*/ ++ ++#define LAST_AVR32_REGNUM 16 ++ ++ ++/** Order of Allocation of Registers **/ ++ ++/* ++If defined, an initializer for a vector of integers, containing the ++numbers of hard registers in the order in which GCC should prefer ++to use them (from most preferred to least). ++ ++If this macro is not defined, registers are used lowest numbered first ++(all else being equal). ++ ++One use of this macro is on machines where the highest numbered ++registers must always be saved and the save-multiple-registers ++instruction supports only sequences of consecutive registers. On such ++machines, define REG_ALLOC_ORDER to be an initializer that lists ++the highest numbered allocable register first. ++*/ ++#define REG_ALLOC_ORDER \ ++{ \ ++ INTERNAL_REGNUM(8), \ ++ INTERNAL_REGNUM(9), \ ++ INTERNAL_REGNUM(10), \ ++ INTERNAL_REGNUM(11), \ ++ INTERNAL_REGNUM(12), \ ++ LR_REGNUM, \ ++ INTERNAL_REGNUM(7), \ ++ INTERNAL_REGNUM(6), \ ++ INTERNAL_REGNUM(5), \ ++ INTERNAL_REGNUM(4), \ ++ INTERNAL_REGNUM(3), \ ++ INTERNAL_REGNUM(2), \ ++ INTERNAL_REGNUM(1), \ ++ INTERNAL_REGNUM(0), \ ++ INTERNAL_FP_REGNUM(15), \ ++ INTERNAL_FP_REGNUM(14), \ ++ INTERNAL_FP_REGNUM(13), \ ++ INTERNAL_FP_REGNUM(12), \ ++ INTERNAL_FP_REGNUM(11), \ ++ INTERNAL_FP_REGNUM(10), \ ++ INTERNAL_FP_REGNUM(9), \ ++ INTERNAL_FP_REGNUM(8), \ ++ INTERNAL_FP_REGNUM(7), \ ++ INTERNAL_FP_REGNUM(6), \ ++ INTERNAL_FP_REGNUM(5), \ ++ INTERNAL_FP_REGNUM(4), \ ++ INTERNAL_FP_REGNUM(3), \ ++ INTERNAL_FP_REGNUM(2), \ ++ INTERNAL_FP_REGNUM(1), \ ++ INTERNAL_FP_REGNUM(0), \ ++ SP_REGNUM, \ ++ PC_REGNUM \ ++} ++ ++ ++/** How Values Fit in Registers **/ ++ ++/* ++A C expression for the number of consecutive hard registers, starting ++at register number REGNO, required to hold a value of mode ++MODE. ++ ++On a machine where all registers are exactly one word, a suitable ++definition of this macro is ++ ++#define HARD_REGNO_NREGS(REGNO, MODE) \ ++ ((GET_MODE_SIZE (MODE) + UNITS_PER_WORD - 1) \ ++ / UNITS_PER_WORD) ++*/ ++#define HARD_REGNO_NREGS(REGNO, MODE) \ ++ ((unsigned int)((GET_MODE_SIZE(MODE) + UNITS_PER_WORD -1 ) / UNITS_PER_WORD)) ++ ++/* ++A C expression that is nonzero if it is permissible to store a value ++of mode MODE in hard register number REGNO (or in several ++registers starting with that one). For a machine where all registers ++are equivalent, a suitable definition is ++ ++ #define HARD_REGNO_MODE_OK(REGNO, MODE) 1 ++ ++You need not include code to check for the numbers of fixed registers, ++because the allocation mechanism considers them to be always occupied. ++ ++On some machines, double-precision values must be kept in even/odd ++register pairs. You can implement that by defining this macro to reject ++odd register numbers for such modes. ++ ++The minimum requirement for a mode to be OK in a register is that the ++mov[mode] instruction pattern support moves between the ++register and other hard register in the same class and that moving a ++value into the register and back out not alter it. ++ ++Since the same instruction used to move word_mode will work for ++all narrower integer modes, it is not necessary on any machine for ++HARD_REGNO_MODE_OK to distinguish between these modes, provided ++you define patterns movhi, etc., to take advantage of this. This ++is useful because of the interaction between HARD_REGNO_MODE_OK ++and MODES_TIEABLE_P; it is very desirable for all integer modes ++to be tieable. ++ ++Many machines have special registers for floating point arithmetic. ++Often people assume that floating point machine modes are allowed only ++in floating point registers. This is not true. Any registers that ++can hold integers can safely hold a floating point machine ++mode, whether or not floating arithmetic can be done on it in those ++registers. Integer move instructions can be used to move the values. ++ ++On some machines, though, the converse is true: fixed-point machine ++modes may not go in floating registers. This is true if the floating ++registers normalize any value stored in them, because storing a ++non-floating value there would garble it. In this case, ++HARD_REGNO_MODE_OK should reject fixed-point machine modes in ++floating registers. But if the floating registers do not automatically ++normalize, if you can store any bit pattern in one and retrieve it ++unchanged without a trap, then any machine mode may go in a floating ++register, so you can define this macro to say so. ++ ++The primary significance of special floating registers is rather that ++they are the registers acceptable in floating point arithmetic ++instructions. However, this is of no concern to ++HARD_REGNO_MODE_OK. You handle it by writing the proper ++constraints for those instructions. ++ ++On some machines, the floating registers are especially slow to access, ++so that it is better to store a value in a stack frame than in such a ++register if floating point arithmetic is not being done. As long as the ++floating registers are not in class GENERAL_REGS, they will not ++be used unless some pattern's constraint asks for one. ++*/ ++#define HARD_REGNO_MODE_OK(REGNO, MODE) avr32_hard_regno_mode_ok(REGNO, MODE) ++ ++/* ++A C expression that is nonzero if a value of mode ++MODE1 is accessible in mode MODE2 without copying. ++ ++If HARD_REGNO_MODE_OK(R, MODE1) and ++HARD_REGNO_MODE_OK(R, MODE2) are always the same for ++any R, then MODES_TIEABLE_P(MODE1, MODE2) ++should be nonzero. If they differ for any R, you should define ++this macro to return zero unless some other mechanism ensures the ++accessibility of the value in a narrower mode. ++ ++You should define this macro to return nonzero in as many cases as ++possible since doing so will allow GCC to perform better register ++allocation. ++*/ ++#define MODES_TIEABLE_P(MODE1, MODE2) \ ++ (GET_MODE_CLASS (MODE1) == GET_MODE_CLASS (MODE2)) ++ ++ ++ ++/****************************************************************************** ++ * Register Classes ++ *****************************************************************************/ ++ ++/* ++An enumeral type that must be defined with all the register class names ++as enumeral values. NO_REGS must be first. ALL_REGS ++must be the last register class, followed by one more enumeral value, ++LIM_REG_CLASSES, which is not a register class but rather ++tells how many classes there are. ++ ++Each register class has a number, which is the value of casting ++the class name to type int. The number serves as an index ++in many of the tables described below. ++*/ ++enum reg_class ++{ ++ NO_REGS, ++ GENERAL_REGS, ++ FP_REGS, ++ ALL_REGS, ++ LIM_REG_CLASSES ++}; ++ ++/* ++The number of distinct register classes, defined as follows: ++ #define N_REG_CLASSES (int) LIM_REG_CLASSES ++*/ ++#define N_REG_CLASSES (int)LIM_REG_CLASSES ++ ++/* ++An initializer containing the names of the register classes as C string ++constants. These names are used in writing some of the debugging dumps. ++*/ ++#define REG_CLASS_NAMES \ ++{ \ ++ "NO_REGS", \ ++ "GENERAL_REGS", \ ++ "FLOATING_POINT_REGS", \ ++ "ALL_REGS" \ ++} ++ ++/* ++An initializer containing the contents of the register classes, as integers ++which are bit masks. The nth integer specifies the contents of class ++n. The way the integer mask is interpreted is that ++register r is in the class if mask & (1 << r) is 1. ++ ++When the machine has more than 32 registers, an integer does not suffice. ++Then the integers are replaced by sub-initializers, braced groupings containing ++several integers. Each sub-initializer must be suitable as an initializer ++for the type HARD_REG_SET which is defined in hard-reg-set.h. ++In this situation, the first integer in each sub-initializer corresponds to ++registers 0 through 31, the second integer to registers 32 through 63, and ++so on. ++*/ ++#define REG_CLASS_CONTENTS { \ ++ {0x00000000}, /* NO_REGS */ \ ++ {0x0000FFFF}, /* GENERAL_REGS */ \ ++ {0xFFFF0000}, /* FP_REGS */ \ ++ {0x7FFFFFFF}, /* ALL_REGS */ \ ++} ++ ++ ++/* ++A C expression whose value is a register class containing hard register ++REGNO. In general there is more than one such class; choose a class ++which is minimal, meaning that no smaller class also contains the ++register. ++*/ ++#define REGNO_REG_CLASS(REGNO) ((REGNO < 16) ? GENERAL_REGS : FP_REGS) ++ ++/* ++A macro whose definition is the name of the class to which a valid ++base register must belong. A base register is one used in an address ++which is the register value plus a displacement. ++*/ ++#define BASE_REG_CLASS GENERAL_REGS ++ ++/* ++This is a variation of the BASE_REG_CLASS macro which allows ++the selection of a base register in a mode depenedent manner. If ++mode is VOIDmode then it should return the same value as ++BASE_REG_CLASS. ++*/ ++#define MODE_BASE_REG_CLASS(MODE) BASE_REG_CLASS ++ ++/* ++A macro whose definition is the name of the class to which a valid ++index register must belong. An index register is one used in an ++address where its value is either multiplied by a scale factor or ++added to another register (as well as added to a displacement). ++*/ ++#define INDEX_REG_CLASS BASE_REG_CLASS ++ ++/* ++A C expression which defines the machine-dependent operand constraint ++letters for register classes. If CHAR is such a letter, the ++value should be the register class corresponding to it. Otherwise, ++the value should be NO_REGS. The register letter r, ++corresponding to class GENERAL_REGS, will not be passed ++to this macro; you do not need to handle it. ++*/ ++#define REG_CLASS_FROM_LETTER(CHAR) ((CHAR) == 'f' ? FP_REGS : NO_REGS) ++ ++ ++/* These assume that REGNO is a hard or pseudo reg number. ++ They give nonzero only if REGNO is a hard reg of the suitable class ++ or a pseudo reg currently allocated to a suitable hard reg. ++ Since they use reg_renumber, they are safe only once reg_renumber ++ has been allocated, which happens in local-alloc.c. */ ++#define TEST_REGNO(R, TEST, VALUE) \ ++ ((R TEST VALUE) || ((unsigned) reg_renumber[R] TEST VALUE)) ++ ++/* ++A C expression which is nonzero if register number num is suitable for use as a base ++register in operand addresses. It may be either a suitable hard register or a pseudo ++register that has been allocated such a hard register. ++*/ ++#define REGNO_OK_FOR_BASE_P(NUM) TEST_REGNO(NUM, <=, LAST_REGNUM) ++ ++/* ++A C expression which is nonzero if register number NUM is ++suitable for use as an index register in operand addresses. It may be ++either a suitable hard register or a pseudo register that has been ++allocated such a hard register. ++ ++The difference between an index register and a base register is that ++the index register may be scaled. If an address involves the sum of ++two registers, neither one of them scaled, then either one may be ++labeled the ``base'' and the other the ``index''; but whichever ++labeling is used must fit the machine's constraints of which registers ++may serve in each capacity. The compiler will try both labelings, ++looking for one that is valid, and will reload one or both registers ++only if neither labeling works. ++*/ ++#define REGNO_OK_FOR_INDEX_P(NUM) TEST_REGNO(NUM, <=, LAST_REGNUM) ++ ++/* ++A C expression that places additional restrictions on the register class ++to use when it is necessary to copy value X into a register in class ++CLASS. The value is a register class; perhaps CLASS, or perhaps ++another, smaller class. On many machines, the following definition is ++safe: #define PREFERRED_RELOAD_CLASS(X,CLASS) CLASS ++ ++Sometimes returning a more restrictive class makes better code. For ++example, on the 68000, when X is an integer constant that is in range ++for a 'moveq' instruction, the value of this macro is always ++DATA_REGS as long as CLASS includes the data registers. ++Requiring a data register guarantees that a 'moveq' will be used. ++ ++If X is a const_double, by returning NO_REGS ++you can force X into a memory constant. This is useful on ++certain machines where immediate floating values cannot be loaded into ++certain kinds of registers. ++*/ ++#define PREFERRED_RELOAD_CLASS(X, CLASS) CLASS ++ ++ ++ ++/* ++A C expression for the maximum number of consecutive registers ++of class CLASS needed to hold a value of mode MODE. ++ ++This is closely related to the macro HARD_REGNO_NREGS. In fact, ++the value of the macro CLASS_MAX_NREGS(CLASS, MODE) ++should be the maximum value of HARD_REGNO_NREGS(REGNO, MODE) ++for all REGNO values in the class CLASS. ++ ++This macro helps control the handling of multiple-word values ++in the reload pass. ++*/ ++#define CLASS_MAX_NREGS(CLASS, MODE) /* ToDo:fixme */ \ ++ (unsigned int)((GET_MODE_SIZE(MODE) + UNITS_PER_WORD - 1) / UNITS_PER_WORD) ++ ++ ++/* ++ Using CONST_OK_FOR_CONSTRAINT_P instead of CONS_OK_FOR_LETTER_P ++ in order to support constraints with more than one letter. ++ Only two letters are then used for constant constraints, ++ the letter 'K' and the letter 'I'. The constraint starting with ++ these letters must consist of four characters. The character following ++ 'K' or 'I' must be either 'u' (unsigned) or 's' (signed) to specify ++ if the constant is zero or sign extended. The last two characters specify ++ the length in bits of the constant. The base constraint letter 'I' means ++ that this is an negated constant, meaning that actually -VAL should be ++ checked to lie withing the valid range instead of VAL which is used when ++ 'K' is the base constraint letter. ++ ++*/ ++ ++#define CONSTRAINT_LEN(C, STR) \ ++ ( ((C) == 'K' || (C) == 'I') ? 4 : \ ++ ((C) == 'R') ? 5 : \ ++ ((C) == 'N' || (C) == 'O' || \ ++ (C) == 'P' || (C) == 'L' || (C) == 'J') ? -1 : \ ++ DEFAULT_CONSTRAINT_LEN((C), (STR)) ) ++ ++#define CONST_OK_FOR_CONSTRAINT_P(VALUE, C, STR) \ ++ avr32_const_ok_for_constraint_p(VALUE, C, STR) ++ ++/* ++A C expression that defines the machine-dependent operand constraint ++letters that specify particular ranges of const_double values ('G' or 'H'). ++ ++If C is one of those letters, the expression should check that ++VALUE, an RTX of code const_double, is in the appropriate ++range and return 1 if so, 0 otherwise. If C is not one of those ++letters, the value should be 0 regardless of VALUE. ++ ++const_double is used for all floating-point constants and for ++DImode fixed-point constants. A given letter can accept either ++or both kinds of values. It can use GET_MODE to distinguish ++between these kinds. ++*/ ++#define CONST_DOUBLE_OK_FOR_LETTER_P(OP, C) \ ++ ((C) == 'G' ? avr32_const_double_immediate(OP) : 0) ++ ++/* ++A C expression that defines the optional machine-dependent constraint ++letters that can be used to segregate specific types of operands, usually ++memory references, for the target machine. Any letter that is not ++elsewhere defined and not matched by REG_CLASS_FROM_LETTER ++may be used. Normally this macro will not be defined. ++ ++If it is required for a particular target machine, it should return 1 ++if VALUE corresponds to the operand type represented by the ++constraint letter C. If C is not defined as an extra ++constraint, the value returned should be 0 regardless of VALUE. ++ ++For example, on the ROMP, load instructions cannot have their output ++in r0 if the memory reference contains a symbolic address. Constraint ++letter 'Q' is defined as representing a memory address that does ++not contain a symbolic address. An alternative is specified with ++a 'Q' constraint on the input and 'r' on the output. The next ++alternative specifies 'm' on the input and a register class that ++does not include r0 on the output. ++*/ ++#define EXTRA_CONSTRAINT_STR(OP, C, STR) \ ++ ((C) == 'W' ? avr32_address_operand(OP, GET_MODE(OP)) : \ ++ (C) == 'R' ? (avr32_indirect_register_operand(OP, GET_MODE(OP)) || \ ++ (avr32_imm_disp_memory_operand(OP, GET_MODE(OP)) \ ++ && avr32_const_ok_for_constraint_p( \ ++ INTVAL(XEXP(XEXP(OP, 0), 1)), \ ++ (STR)[1], &(STR)[1]))) : \ ++ (C) == 'S' ? avr32_indexed_memory_operand(OP, GET_MODE(OP)) : \ ++ (C) == 'T' ? avr32_const_pool_ref_operand(OP, GET_MODE(OP)) : \ ++ (C) == 'U' ? SYMBOL_REF_RCALL_FUNCTION_P(OP) : \ ++ (C) == 'Z' ? avr32_cop_memory_operand(OP, GET_MODE(OP)) : \ ++ 0) ++ ++ ++#define EXTRA_MEMORY_CONSTRAINT(C, STR) ( ((C) == 'R') || \ ++ ((C) == 'S') || \ ++ ((C) == 'Z') ) ++ ++ ++/* Returns nonzero if op is a function SYMBOL_REF which ++ can be called using an rcall instruction */ ++#define SYMBOL_REF_RCALL_FUNCTION_P(op) \ ++ ( GET_CODE(op) == SYMBOL_REF \ ++ && SYMBOL_REF_FUNCTION_P(op) \ ++ && SYMBOL_REF_LOCAL_P(op) \ ++ && !SYMBOL_REF_EXTERNAL_P(op) \ ++ && !TARGET_HAS_ASM_ADDR_PSEUDOS ) ++ ++/****************************************************************************** ++ * Stack Layout and Calling Conventions ++ *****************************************************************************/ ++ ++/** Basic Stack Layout **/ ++ ++/* ++Define this macro if pushing a word onto the stack moves the stack ++pointer to a smaller address. ++ ++When we say, ``define this macro if ...,'' it means that the ++compiler checks this macro only with #ifdef so the precise ++definition used does not matter. ++*/ ++/* pushm decrece SP: *(--SP) <-- Rx */ ++#define STACK_GROWS_DOWNWARD ++ ++/* ++This macro defines the operation used when something is pushed ++on the stack. In RTL, a push operation will be ++(set (mem (STACK_PUSH_CODE (reg sp))) ...) ++ ++The choices are PRE_DEC, POST_DEC, PRE_INC, ++and POST_INC. Which of these is correct depends on ++the stack direction and on whether the stack pointer points ++to the last item on the stack or whether it points to the ++space for the next item on the stack. ++ ++The default is PRE_DEC when STACK_GROWS_DOWNWARD is ++defined, which is almost always right, and PRE_INC otherwise, ++which is often wrong. ++*/ ++/* pushm: *(--SP) <-- Rx */ ++#define STACK_PUSH_CODE PRE_DEC ++ ++/* Define this to nonzero if the nominal address of the stack frame ++ is at the high-address end of the local variables; ++ that is, each additional local variable allocated ++ goes at a more negative offset in the frame. */ ++#define FRAME_GROWS_DOWNWARD 1 ++ ++ ++/* ++Offset from the frame pointer to the first local variable slot to be allocated. ++ ++If FRAME_GROWS_DOWNWARD, find the next slot's offset by ++subtracting the first slot's length from STARTING_FRAME_OFFSET. ++Otherwise, it is found by adding the length of the first slot to the ++value STARTING_FRAME_OFFSET. ++ (i'm not sure if the above is still correct.. had to change it to get ++ rid of an overfull. --mew 2feb93 ) ++*/ ++#define STARTING_FRAME_OFFSET 0 ++ ++/* ++Offset from the stack pointer register to the first location at which ++outgoing arguments are placed. If not specified, the default value of ++zero is used. This is the proper value for most machines. ++ ++If ARGS_GROW_DOWNWARD, this is the offset to the location above ++the first location at which outgoing arguments are placed. ++*/ ++#define STACK_POINTER_OFFSET 0 ++ ++/* ++Offset from the argument pointer register to the first argument's ++address. On some machines it may depend on the data type of the ++function. ++ ++If ARGS_GROW_DOWNWARD, this is the offset to the location above ++the first argument's address. ++*/ ++#define FIRST_PARM_OFFSET(FUNDECL) 0 ++ ++ ++/* ++A C expression whose value is RTL representing the address in a stack ++frame where the pointer to the caller's frame is stored. Assume that ++FRAMEADDR is an RTL expression for the address of the stack frame ++itself. ++ ++If you don't define this macro, the default is to return the value ++of FRAMEADDR - that is, the stack frame address is also the ++address of the stack word that points to the previous frame. ++*/ ++#define DYNAMIC_CHAIN_ADDRESS(FRAMEADDR) plus_constant ((FRAMEADDR), 4) ++ ++ ++/* ++A C expression whose value is RTL representing the value of the return ++address for the frame COUNT steps up from the current frame, after ++the prologue. FRAMEADDR is the frame pointer of the COUNT ++frame, or the frame pointer of the COUNT - 1 frame if ++RETURN_ADDR_IN_PREVIOUS_FRAME is defined. ++ ++The value of the expression must always be the correct address when ++COUNT is zero, but may be NULL_RTX if there is not way to ++determine the return address of other frames. ++*/ ++#define RETURN_ADDR_RTX(COUNT, FRAMEADDR) avr32_return_addr(COUNT, FRAMEADDR) ++ ++ ++/* ++A C expression whose value is RTL representing the location of the ++incoming return address at the beginning of any function, before the ++prologue. This RTL is either a REG, indicating that the return ++value is saved in 'REG', or a MEM representing a location in ++the stack. ++ ++You only need to define this macro if you want to support call frame ++debugging information like that provided by DWARF 2. ++ ++If this RTL is a REG, you should also define ++DWARF_FRAME_RETURN_COLUMN to DWARF_FRAME_REGNUM (REGNO). ++*/ ++#define INCOMING_RETURN_ADDR_RTX gen_rtx_REG (Pmode, LR_REGNUM) ++ ++ ++ ++/* ++A C expression whose value is an integer giving the offset, in bytes, ++from the value of the stack pointer register to the top of the stack ++frame at the beginning of any function, before the prologue. The top of ++the frame is defined to be the value of the stack pointer in the ++previous frame, just before the call instruction. ++ ++You only need to define this macro if you want to support call frame ++debugging information like that provided by DWARF 2. ++*/ ++#define INCOMING_FRAME_SP_OFFSET 0 ++ ++ ++/** Exception Handling Support **/ ++ ++/* Use setjump/longjump for exception handling. */ ++#define DWARF2_UNWIND_INFO 0 ++#define MUST_USE_SJLJ_EXCEPTIONS 1 ++ ++/* ++A C expression whose value is the Nth register number used for ++data by exception handlers, or INVALID_REGNUM if fewer than ++N registers are usable. ++ ++The exception handling library routines communicate with the exception ++handlers via a set of agreed upon registers. Ideally these registers ++should be call-clobbered; it is possible to use call-saved registers, ++but may negatively impact code size. The target must support at least ++2 data registers, but should define 4 if there are enough free registers. ++ ++You must define this macro if you want to support call frame exception ++handling like that provided by DWARF 2. ++*/ ++/* ++ Use r9-r11 ++*/ ++#define EH_RETURN_DATA_REGNO(N) \ ++ ((N<3) ? INTERNAL_REGNUM(N+9) : INVALID_REGNUM) ++ ++/* ++A C expression whose value is RTL representing a location in which ++to store a stack adjustment to be applied before function return. ++This is used to unwind the stack to an exception handler's call frame. ++It will be assigned zero on code paths that return normally. ++ ++Typically this is a call-clobbered hard register that is otherwise ++untouched by the epilogue, but could also be a stack slot. ++ ++You must define this macro if you want to support call frame exception ++handling like that provided by DWARF 2. ++*/ ++/* ++ Use r8 ++*/ ++#define EH_RETURN_STACKADJ_REGNO INTERNAL_REGNUM(8) ++#define EH_RETURN_STACKADJ_RTX gen_rtx_REG(SImode, EH_RETURN_STACKADJ_REGNO) ++ ++/* ++A C expression whose value is RTL representing a location in which ++to store the address of an exception handler to which we should ++return. It will not be assigned on code paths that return normally. ++ ++Typically this is the location in the call frame at which the normal ++return address is stored. For targets that return by popping an ++address off the stack, this might be a memory address just below ++the target call frame rather than inside the current call ++frame. EH_RETURN_STACKADJ_RTX will have already been assigned, ++so it may be used to calculate the location of the target call frame. ++ ++Some targets have more complex requirements than storing to an ++address calculable during initial code generation. In that case ++the eh_return instruction pattern should be used instead. ++ ++If you want to support call frame exception handling, you must ++define either this macro or the eh_return instruction pattern. ++*/ ++/* ++ We define the eh_return instruction pattern, so this isn't needed. ++*/ ++/* #define EH_RETURN_HANDLER_RTX gen_rtx_REG(Pmode, RET_REGISTER) */ ++ ++/* ++ This macro chooses the encoding of pointers embedded in the ++ exception handling sections. If at all possible, this should be ++ defined such that the exception handling section will not require ++ dynamic relocations, and so may be read-only. ++ ++ code is 0 for data, 1 for code labels, 2 for function ++ pointers. global is true if the symbol may be affected by dynamic ++ relocations. The macro should return a combination of the DW_EH_PE_* ++ defines as found in dwarf2.h. ++ ++ If this macro is not defined, pointers will not be encoded but ++ represented directly. ++*/ ++#define ASM_PREFERRED_EH_DATA_FORMAT(CODE, GLOBAL) \ ++ ((flag_pic && (GLOBAL) ? DW_EH_PE_indirect : 0) \ ++ | (flag_pic ? DW_EH_PE_pcrel : DW_EH_PE_absptr) \ ++ | DW_EH_PE_sdata4) ++ ++/* ToDo: The rest of this subsection */ ++ ++/** Specifying How Stack Checking is Done **/ ++/* ToDo: All in this subsection */ ++ ++/** Registers That Address the Stack Frame **/ ++ ++/* ++The register number of the stack pointer register, which must also be a ++fixed register according to FIXED_REGISTERS. On most machines, ++the hardware determines which register this is. ++*/ ++/* Using r13 as stack pointer. */ ++#define STACK_POINTER_REGNUM INTERNAL_REGNUM(13) ++ ++/* ++The register number of the frame pointer register, which is used to ++access automatic variables in the stack frame. On some machines, the ++hardware determines which register this is. On other machines, you can ++choose any register you wish for this purpose. ++*/ ++/* Use r7 */ ++#define FRAME_POINTER_REGNUM INTERNAL_REGNUM(7) ++ ++ ++ ++/* ++The register number of the arg pointer register, which is used to access ++the function's argument list. On some machines, this is the same as the ++frame pointer register. On some machines, the hardware determines which ++register this is. On other machines, you can choose any register you ++wish for this purpose. If this is not the same register as the frame ++pointer register, then you must mark it as a fixed register according to ++FIXED_REGISTERS, or arrange to be able to eliminate it (see Section ++10.10.5 [Elimination], page 224). ++*/ ++/* Using r5 */ ++#define ARG_POINTER_REGNUM INTERNAL_REGNUM(4) ++ ++ ++/* ++Register numbers used for passing a function's static chain pointer. If ++register windows are used, the register number as seen by the called ++function is STATIC_CHAIN_INCOMING_REGNUM, while the register ++number as seen by the calling function is STATIC_CHAIN_REGNUM. If ++these registers are the same, STATIC_CHAIN_INCOMING_REGNUM need ++not be defined. ++ ++The static chain register need not be a fixed register. ++ ++If the static chain is passed in memory, these macros should not be ++defined; instead, the next two macros should be defined. ++*/ ++/* Using r0 */ ++#define STATIC_CHAIN_REGNUM INTERNAL_REGNUM(0) ++ ++ ++/** Eliminating Frame Pointer and Arg Pointer **/ ++ ++/* ++A C expression which is nonzero if a function must have and use a frame ++pointer. This expression is evaluated in the reload pass. If its value is ++nonzero the function will have a frame pointer. ++ ++The expression can in principle examine the current function and decide ++according to the facts, but on most machines the constant 0 or the ++constant 1 suffices. Use 0 when the machine allows code to be generated ++with no frame pointer, and doing so saves some time or space. Use 1 ++when there is no possible advantage to avoiding a frame pointer. ++ ++In certain cases, the compiler does not know how to produce valid code ++without a frame pointer. The compiler recognizes those cases and ++automatically gives the function a frame pointer regardless of what ++FRAME_POINTER_REQUIRED says. You don't need to worry about ++them. ++ ++In a function that does not require a frame pointer, the frame pointer ++register can be allocated for ordinary usage, unless you mark it as a ++fixed register. See FIXED_REGISTERS for more information. ++*/ ++/* We need the frame pointer when compiling for profiling */ ++#define FRAME_POINTER_REQUIRED (current_function_profile) ++ ++/* ++A C statement to store in the variable DEPTH_VAR the difference ++between the frame pointer and the stack pointer values immediately after ++the function prologue. The value would be computed from information ++such as the result of get_frame_size () and the tables of ++registers regs_ever_live and call_used_regs. ++ ++If ELIMINABLE_REGS is defined, this macro will be not be used and ++need not be defined. Otherwise, it must be defined even if ++FRAME_POINTER_REQUIRED is defined to always be true; in that ++case, you may set DEPTH_VAR to anything. ++*/ ++#define INITIAL_FRAME_POINTER_OFFSET(DEPTH_VAR) ((DEPTH_VAR) = get_frame_size()) ++ ++/* ++If defined, this macro specifies a table of register pairs used to ++eliminate unneeded registers that point into the stack frame. If it is not ++defined, the only elimination attempted by the compiler is to replace ++references to the frame pointer with references to the stack pointer. ++ ++The definition of this macro is a list of structure initializations, each ++of which specifies an original and replacement register. ++ ++On some machines, the position of the argument pointer is not known until ++the compilation is completed. In such a case, a separate hard register ++must be used for the argument pointer. This register can be eliminated by ++replacing it with either the frame pointer or the argument pointer, ++depending on whether or not the frame pointer has been eliminated. ++ ++In this case, you might specify: ++ #define ELIMINABLE_REGS \ ++ {{ARG_POINTER_REGNUM, STACK_POINTER_REGNUM}, \ ++ {ARG_POINTER_REGNUM, FRAME_POINTER_REGNUM}, \ ++ {FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM}} ++ ++Note that the elimination of the argument pointer with the stack pointer is ++specified first since that is the preferred elimination. ++*/ ++#define ELIMINABLE_REGS \ ++{ \ ++ { FRAME_POINTER_REGNUM, STACK_POINTER_REGNUM }, \ ++ { ARG_POINTER_REGNUM, STACK_POINTER_REGNUM }, \ ++ { ARG_POINTER_REGNUM, FRAME_POINTER_REGNUM } \ ++} ++ ++/* ++A C expression that returns nonzero if the compiler is allowed to try ++to replace register number FROM with register number ++TO. This macro need only be defined if ELIMINABLE_REGS ++is defined, and will usually be the constant 1, since most of the cases ++preventing register elimination are things that the compiler already ++knows about. ++*/ ++#define CAN_ELIMINATE(FROM, TO) 1 ++ ++/* ++This macro is similar to INITIAL_FRAME_POINTER_OFFSET. It ++specifies the initial difference between the specified pair of ++registers. This macro must be defined if ELIMINABLE_REGS is ++defined. ++*/ ++#define INITIAL_ELIMINATION_OFFSET(FROM, TO, OFFSET) \ ++ ((OFFSET) = avr32_initial_elimination_offset(FROM, TO)) ++ ++/** Passing Function Arguments on the Stack **/ ++ ++ ++/* ++A C expression. If nonzero, push insns will be used to pass ++outgoing arguments. ++If the target machine does not have a push instruction, set it to zero. ++That directs GCC to use an alternate strategy: to ++allocate the entire argument block and then store the arguments into ++it. When PUSH_ARGS is nonzero, PUSH_ROUNDING must be defined too. ++*/ ++#define PUSH_ARGS 1 ++ ++ ++/* ++A C expression that is the number of bytes actually pushed onto the ++stack when an instruction attempts to push NPUSHED bytes. ++ ++On some machines, the definition ++ ++ #define PUSH_ROUNDING(BYTES) (BYTES) ++ ++will suffice. But on other machines, instructions that appear ++to push one byte actually push two bytes in an attempt to maintain ++alignment. Then the definition should be ++ ++ #define PUSH_ROUNDING(BYTES) (((BYTES) + 1) & ~1) ++*/ ++/* Push 4 bytes at the time. */ ++#define PUSH_ROUNDING(NPUSHED) (((NPUSHED) + 3) & ~3) ++ ++/* ++A C expression. If nonzero, the maximum amount of space required for ++outgoing arguments will be computed and placed into the variable ++current_function_outgoing_args_size. No space will be pushed ++onto the stack for each call; instead, the function prologue should ++increase the stack frame size by this amount. ++ ++Setting both PUSH_ARGS and ACCUMULATE_OUTGOING_ARGS is not proper. ++*/ ++#define ACCUMULATE_OUTGOING_ARGS 0 ++ ++ ++ ++ ++/* ++A C expression that should indicate the number of bytes of its own ++arguments that a function pops on returning, or 0 if the ++function pops no arguments and the caller must therefore pop them all ++after the function returns. ++ ++FUNDECL is a C variable whose value is a tree node that describes ++the function in question. Normally it is a node of type ++FUNCTION_DECL that describes the declaration of the function. ++From this you can obtain the DECL_ATTRIBUTES of the function. ++ ++FUNTYPE is a C variable whose value is a tree node that ++describes the function in question. Normally it is a node of type ++FUNCTION_TYPE that describes the data type of the function. ++From this it is possible to obtain the data types of the value and ++arguments (if known). ++ ++When a call to a library function is being considered, FUNDECL ++will contain an identifier node for the library function. Thus, if ++you need to distinguish among various library functions, you can do so ++by their names. Note that ``library function'' in this context means ++a function used to perform arithmetic, whose name is known specially ++in the compiler and was not mentioned in the C code being compiled. ++ ++STACK_SIZE is the number of bytes of arguments passed on the ++stack. If a variable number of bytes is passed, it is zero, and ++argument popping will always be the responsibility of the calling function. ++ ++On the VAX, all functions always pop their arguments, so the definition ++of this macro is STACK_SIZE. On the 68000, using the standard ++calling convention, no functions pop their arguments, so the value of ++the macro is always 0 in this case. But an alternative calling ++convention is available in which functions that take a fixed number of ++arguments pop them but other functions (such as printf) pop ++nothing (the caller pops all). When this convention is in use, ++FUNTYPE is examined to determine whether a function takes a fixed ++number of arguments. ++*/ ++#define RETURN_POPS_ARGS(FUNDECL, FUNTYPE, STACK_SIZE) 0 ++ ++ ++/*Return true if this function can we use a single return instruction*/ ++#define USE_RETURN_INSN(ISCOND) avr32_use_return_insn(ISCOND) ++ ++/* ++A C expression that should indicate the number of bytes a call sequence ++pops off the stack. It is added to the value of RETURN_POPS_ARGS ++when compiling a function call. ++ ++CUM is the variable in which all arguments to the called function ++have been accumulated. ++ ++On certain architectures, such as the SH5, a call trampoline is used ++that pops certain registers off the stack, depending on the arguments ++that have been passed to the function. Since this is a property of the ++call site, not of the called function, RETURN_POPS_ARGS is not ++appropriate. ++*/ ++#define CALL_POPS_ARGS(CUM) 0 ++ ++/* Passing Arguments in Registers */ ++ ++/* ++A C expression that controls whether a function argument is passed ++in a register, and which register. ++ ++The arguments are CUM, which summarizes all the previous ++arguments; MODE, the machine mode of the argument; TYPE, ++the data type of the argument as a tree node or 0 if that is not known ++(which happens for C support library functions); and NAMED, ++which is 1 for an ordinary argument and 0 for nameless arguments that ++correspond to '...' in the called function's prototype. ++TYPE can be an incomplete type if a syntax error has previously ++occurred. ++ ++The value of the expression is usually either a reg RTX for the ++hard register in which to pass the argument, or zero to pass the ++argument on the stack. ++ ++For machines like the VAX and 68000, where normally all arguments are ++pushed, zero suffices as a definition. ++ ++The value of the expression can also be a parallel RTX. This is ++used when an argument is passed in multiple locations. The mode of the ++of the parallel should be the mode of the entire argument. The ++parallel holds any number of expr_list pairs; each one ++describes where part of the argument is passed. In each ++expr_list the first operand must be a reg RTX for the hard ++register in which to pass this part of the argument, and the mode of the ++register RTX indicates how large this part of the argument is. The ++second operand of the expr_list is a const_int which gives ++the offset in bytes into the entire argument of where this part starts. ++As a special exception the first expr_list in the parallel ++RTX may have a first operand of zero. This indicates that the entire ++argument is also stored on the stack. ++ ++The last time this macro is called, it is called with MODE == VOIDmode, ++and its result is passed to the call or call_value ++pattern as operands 2 and 3 respectively. ++ ++The usual way to make the ISO library 'stdarg.h' work on a machine ++where some arguments are usually passed in registers, is to cause ++nameless arguments to be passed on the stack instead. This is done ++by making FUNCTION_ARG return 0 whenever NAMED is 0. ++ ++You may use the macro MUST_PASS_IN_STACK (MODE, TYPE) ++in the definition of this macro to determine if this argument is of a ++type that must be passed in the stack. If REG_PARM_STACK_SPACE ++is not defined and FUNCTION_ARG returns nonzero for such an ++argument, the compiler will abort. If REG_PARM_STACK_SPACE is ++defined, the argument will be computed in the stack and then loaded into ++a register. */ ++ ++#define FUNCTION_ARG(CUM, MODE, TYPE, NAMED) \ ++ avr32_function_arg(&(CUM), MODE, TYPE, NAMED) ++ ++ ++ ++ ++/* ++A C type for declaring a variable that is used as the first argument of ++FUNCTION_ARG and other related values. For some target machines, ++the type int suffices and can hold the number of bytes of ++argument so far. ++ ++There is no need to record in CUMULATIVE_ARGS anything about the ++arguments that have been passed on the stack. The compiler has other ++variables to keep track of that. For target machines on which all ++arguments are passed on the stack, there is no need to store anything in ++CUMULATIVE_ARGS; however, the data structure must exist and ++should not be empty, so use int. ++*/ ++typedef struct avr32_args ++{ ++ /* Index representing the argument register the current function argument ++ will occupy */ ++ int index; ++ /* A mask with bits representing the argument registers: if a bit is set ++ then this register is used for an arguemnt */ ++ int used_index; ++ /* TRUE if this function has anonymous arguments */ ++ int uses_anonymous_args; ++ /* The size in bytes of the named arguments pushed on the stack */ ++ int stack_pushed_args_size; ++ /* Set to true if this function needs a Return Value Pointer */ ++ int use_rvp; ++ ++} CUMULATIVE_ARGS; ++ ++ ++#define FIRST_CUM_REG_INDEX 0 ++#define LAST_CUM_REG_INDEX 4 ++#define GET_REG_INDEX(CUM) ((CUM)->index) ++#define SET_REG_INDEX(CUM, INDEX) ((CUM)->index = (INDEX)); ++#define GET_USED_INDEX(CUM, INDEX) ((CUM)->used_index & (1 << (INDEX))) ++#define SET_USED_INDEX(CUM, INDEX) \ ++ do \ ++ { \ ++ if (INDEX >= 0) \ ++ (CUM)->used_index |= (1 << (INDEX)); \ ++ } \ ++ while (0) ++#define SET_INDEXES_UNUSED(CUM) ((CUM)->used_index = 0) ++ ++ ++/* ++ A C statement (sans semicolon) for initializing the variable cum for the ++ state at the beginning of the argument list. The variable has type ++ CUMULATIVE_ARGS. The value of FNTYPE is the tree node for the data type of ++ the function which will receive the args, or 0 if the args are to a compiler ++ support library function. For direct calls that are not libcalls, FNDECL ++ contain the declaration node of the function. FNDECL is also set when ++ INIT_CUMULATIVE_ARGS is used to find arguments for the function being ++ compiled. N_NAMED_ARGS is set to the number of named arguments, including a ++ structure return address if it is passed as a parameter, when making a call. ++ When processing incoming arguments, N_NAMED_ARGS is set to -1. ++ ++ When processing a call to a compiler support library function, LIBNAME ++ identifies which one. It is a symbol_ref rtx which contains the name of the ++ function, as a string. LIBNAME is 0 when an ordinary C function call is ++ being processed. Thus, each time this macro is called, either LIBNAME or ++ FNTYPE is nonzero, but never both of them at once. ++*/ ++#define INIT_CUMULATIVE_ARGS(CUM, FNTYPE, LIBNAME, FNDECL, N_NAMED_ARGS) \ ++ avr32_init_cumulative_args(&(CUM), FNTYPE, LIBNAME, FNDECL) ++ ++ ++/* ++A C statement (sans semicolon) to update the summarizer variable ++CUM to advance past an argument in the argument list. The ++values MODE, TYPE and NAMED describe that argument. ++Once this is done, the variable CUM is suitable for analyzing ++the following argument with FUNCTION_ARG, etc. ++ ++This macro need not do anything if the argument in question was passed ++on the stack. The compiler knows how to track the amount of stack space ++used for arguments without any special help. ++*/ ++#define FUNCTION_ARG_ADVANCE(CUM, MODE, TYPE, NAMED) \ ++ avr32_function_arg_advance(&(CUM), MODE, TYPE, NAMED) ++ ++/* ++If defined, a C expression which determines whether, and in which direction, ++to pad out an argument with extra space. The value should be of type ++enum direction: either 'upward' to pad above the argument, ++'downward' to pad below, or 'none' to inhibit padding. ++ ++The amount of padding is always just enough to reach the next ++multiple of FUNCTION_ARG_BOUNDARY; this macro does not control ++it. ++ ++This macro has a default definition which is right for most systems. ++For little-endian machines, the default is to pad upward. For ++big-endian machines, the default is to pad downward for an argument of ++constant size shorter than an int, and upward otherwise. ++*/ ++#define FUNCTION_ARG_PADDING(MODE, TYPE) \ ++ avr32_function_arg_padding(MODE, TYPE) ++ ++/* ++ Specify padding for the last element of a block move between registers ++ and memory. First is nonzero if this is the only element. Defining ++ this macro allows better control of register function parameters on ++ big-endian machines, without using PARALLEL rtl. In particular, ++ MUST_PASS_IN_STACK need not test padding and mode of types in registers, ++ as there is no longer a "wrong" part of a register; For example, a three ++ byte aggregate may be passed in the high part of a register if so required. ++*/ ++#define BLOCK_REG_PADDING(MODE, TYPE, FIRST) \ ++ avr32_function_arg_padding(MODE, TYPE) ++ ++/* ++If defined, a C expression which determines whether the default ++implementation of va_arg will attempt to pad down before reading the ++next argument, if that argument is smaller than its aligned space as ++controlled by PARM_BOUNDARY. If this macro is not defined, all such ++arguments are padded down if BYTES_BIG_ENDIAN is true. ++*/ ++#define PAD_VARARGS_DOWN \ ++ (FUNCTION_ARG_PADDING (TYPE_MODE (type), type) == downward) ++ ++ ++/* ++A C expression that is nonzero if REGNO is the number of a hard ++register in which function arguments are sometimes passed. This does ++not include implicit arguments such as the static chain and ++the structure-value address. On many machines, no registers can be ++used for this purpose since all function arguments are pushed on the ++stack. ++*/ ++/* ++ Use r8 - r12 for function arguments. ++*/ ++#define FUNCTION_ARG_REGNO_P(REGNO) \ ++ (REGNO >= 3 && REGNO <= 7) ++ ++/* Number of registers used for passing function arguments */ ++#define NUM_ARG_REGS 5 ++ ++/* ++If defined, the order in which arguments are loaded into their ++respective argument registers is reversed so that the last ++argument is loaded first. This macro only affects arguments ++passed in registers. ++*/ ++/* #define LOAD_ARGS_REVERSED */ ++ ++/** How Scalar Function Values Are Returned **/ ++ ++/* AVR32 is using r12 as return register. */ ++#define RET_REGISTER (15 - 12) ++ ++ ++/* ++A C expression to create an RTX representing the place where a library ++function returns a value of mode MODE. If the precise function ++being called is known, FUNC is a tree node ++(FUNCTION_DECL) for it; otherwise, func is a null ++pointer. This makes it possible to use a different value-returning ++convention for specific functions when all their calls are ++known. ++ ++Note that "library function" in this context means a compiler ++support routine, used to perform arithmetic, whose name is known ++specially by the compiler and was not mentioned in the C code being ++compiled. ++ ++The definition of LIBRARY_VALUE need not be concerned aggregate ++data types, because none of the library functions returns such types. ++*/ ++#define LIBCALL_VALUE(MODE) avr32_libcall_value(MODE) ++ ++/* ++A C expression that is nonzero if REGNO is the number of a hard ++register in which the values of called function may come back. ++ ++A register whose use for returning values is limited to serving as the ++second of a pair (for a value of type double, say) need not be ++recognized by this macro. So for most machines, this definition ++suffices: ++ #define FUNCTION_VALUE_REGNO_P(N) ((N) == 0) ++ ++If the machine has register windows, so that the caller and the called ++function use different registers for the return value, this macro ++should recognize only the caller's register numbers. ++*/ ++/* ++ When returning a value of mode DImode, r11:r10 is used, else r12 is used. ++*/ ++#define FUNCTION_VALUE_REGNO_P(REGNO) ((REGNO) == RET_REGISTER \ ++ || (REGNO) == INTERNAL_REGNUM(11)) ++ ++ ++/** How Large Values Are Returned **/ ++ ++ ++/* ++Define this macro to be 1 if all structure and union return values must be ++in memory. Since this results in slower code, this should be defined ++only if needed for compatibility with other compilers or with an ABI. ++If you define this macro to be 0, then the conventions used for structure ++and union return values are decided by the RETURN_IN_MEMORY macro. ++ ++If not defined, this defaults to the value 1. ++*/ ++#define DEFAULT_PCC_STRUCT_RETURN 0 ++ ++ ++ ++ ++/** Generating Code for Profiling **/ ++ ++/* ++A C statement or compound statement to output to FILE some ++assembler code to call the profiling subroutine mcount. ++ ++The details of how mcount expects to be called are determined by ++your operating system environment, not by GCC. To figure them out, ++compile a small program for profiling using the system's installed C ++compiler and look at the assembler code that results. ++ ++Older implementations of mcount expect the address of a counter ++variable to be loaded into some register. The name of this variable is ++'LP' followed by the number LABELNO, so you would generate ++the name using 'LP%d' in a fprintf. ++*/ ++/* ToDo: fixme */ ++#ifndef FUNCTION_PROFILER ++#define FUNCTION_PROFILER(FILE, LABELNO) \ ++ fprintf((FILE), "/* profiler %d */", (LABELNO)) ++#endif ++ ++ ++/***************************************************************************** ++ * Trampolines for Nested Functions * ++ *****************************************************************************/ ++ ++/* ++A C statement to output, on the stream FILE, assembler code for a ++block of data that contains the constant parts of a trampoline. This ++code should not include a label - the label is taken care of ++automatically. ++ ++If you do not define this macro, it means no template is needed ++for the target. Do not define this macro on systems where the block move ++code to copy the trampoline into place would be larger than the code ++to generate it on the spot. ++*/ ++/* ToDo: correct? */ ++#define TRAMPOLINE_TEMPLATE(FILE) avr32_trampoline_template(FILE); ++ ++ ++/* ++A C expression for the size in bytes of the trampoline, as an integer. ++*/ ++/* ToDo: fixme */ ++#define TRAMPOLINE_SIZE 0x0C ++ ++/* ++Alignment required for trampolines, in bits. ++ ++If you don't define this macro, the value of BIGGEST_ALIGNMENT ++is used for aligning trampolines. ++*/ ++#define TRAMPOLINE_ALIGNMENT 16 ++ ++/* ++A C statement to initialize the variable parts of a trampoline. ++ADDR is an RTX for the address of the trampoline; FNADDR is ++an RTX for the address of the nested function; STATIC_CHAIN is an ++RTX for the static chain value that should be passed to the function ++when it is called. ++*/ ++#define INITIALIZE_TRAMPOLINE(ADDR, FNADDR, STATIC_CHAIN) \ ++ avr32_initialize_trampoline(ADDR, FNADDR, STATIC_CHAIN) ++ ++ ++/****************************************************************************** ++ * Implicit Calls to Library Routines ++ *****************************************************************************/ ++ ++/* Tail calling. */ ++ ++/* A C expression that evaluates to true if it is ok to perform a sibling ++ call to DECL. */ ++#define FUNCTION_OK_FOR_SIBCALL(DECL) 0 ++ ++#define OVERRIDE_OPTIONS avr32_override_options () ++ ++#define OPTIMIZATION_OPTIONS(LEVEL, SIZE) avr32_optimization_options (LEVEL, SIZE) ++ ++ ++/****************************************************************************** ++ * Addressing Modes ++ *****************************************************************************/ ++ ++/* ++A C expression that is nonzero if the machine supports pre-increment, ++pre-decrement, post-increment, or post-decrement addressing respectively. ++*/ ++/* ++ AVR32 supports Rp++ and --Rp ++*/ ++#define HAVE_PRE_INCREMENT 0 ++#define HAVE_PRE_DECREMENT 1 ++#define HAVE_POST_INCREMENT 1 ++#define HAVE_POST_DECREMENT 0 ++ ++/* ++A C expression that is nonzero if the machine supports pre- or ++post-address side-effect generation involving constants other than ++the size of the memory operand. ++*/ ++#define HAVE_PRE_MODIFY_DISP 0 ++#define HAVE_POST_MODIFY_DISP 0 ++ ++/* ++A C expression that is nonzero if the machine supports pre- or ++post-address side-effect generation involving a register displacement. ++*/ ++#define HAVE_PRE_MODIFY_REG 0 ++#define HAVE_POST_MODIFY_REG 0 ++ ++/* ++A C expression that is 1 if the RTX X is a constant which ++is a valid address. On most machines, this can be defined as ++CONSTANT_P (X), but a few machines are more restrictive ++in which constant addresses are supported. ++ ++CONSTANT_P accepts integer-values expressions whose values are ++not explicitly known, such as symbol_ref, label_ref, and ++high expressions and const arithmetic expressions, in ++addition to const_int and const_double expressions. ++*/ ++#define CONSTANT_ADDRESS_P(X) CONSTANT_P(X) ++ ++/* ++A number, the maximum number of registers that can appear in a valid ++memory address. Note that it is up to you to specify a value equal to ++the maximum number that GO_IF_LEGITIMATE_ADDRESS would ever ++accept. ++*/ ++#define MAX_REGS_PER_ADDRESS 2 ++ ++/* ++A C compound statement with a conditional goto LABEL; ++executed if X (an RTX) is a legitimate memory address on the ++target machine for a memory operand of mode MODE. ++ ++It usually pays to define several simpler macros to serve as ++subroutines for this one. Otherwise it may be too complicated to ++understand. ++ ++This macro must exist in two variants: a strict variant and a ++non-strict one. The strict variant is used in the reload pass. It ++must be defined so that any pseudo-register that has not been ++allocated a hard register is considered a memory reference. In ++contexts where some kind of register is required, a pseudo-register ++with no hard register must be rejected. ++ ++The non-strict variant is used in other passes. It must be defined to ++accept all pseudo-registers in every context where some kind of ++register is required. ++ ++Compiler source files that want to use the strict variant of this ++macro define the macro REG_OK_STRICT. You should use an ++#ifdef REG_OK_STRICT conditional to define the strict variant ++in that case and the non-strict variant otherwise. ++ ++Subroutines to check for acceptable registers for various purposes (one ++for base registers, one for index registers, and so on) are typically ++among the subroutines used to define GO_IF_LEGITIMATE_ADDRESS. ++Then only these subroutine macros need have two variants; the higher ++levels of macros may be the same whether strict or not. ++ ++Normally, constant addresses which are the sum of a symbol_ref ++and an integer are stored inside a const RTX to mark them as ++constant. Therefore, there is no need to recognize such sums ++specifically as legitimate addresses. Normally you would simply ++recognize any const as legitimate. ++ ++Usually PRINT_OPERAND_ADDRESS is not prepared to handle constant ++sums that are not marked with const. It assumes that a naked ++plus indicates indexing. If so, then you must reject such ++naked constant sums as illegitimate addresses, so that none of them will ++be given to PRINT_OPERAND_ADDRESS. ++ ++On some machines, whether a symbolic address is legitimate depends on ++the section that the address refers to. On these machines, define the ++macro ENCODE_SECTION_INFO to store the information into the ++symbol_ref, and then check for it here. When you see a ++const, you will have to look inside it to find the ++symbol_ref in order to determine the section. ++ ++The best way to modify the name string is by adding text to the ++beginning, with suitable punctuation to prevent any ambiguity. Allocate ++the new name in saveable_obstack. You will have to modify ++ASM_OUTPUT_LABELREF to remove and decode the added text and ++output the name accordingly, and define STRIP_NAME_ENCODING to ++access the original name string. ++ ++You can check the information stored here into the symbol_ref in ++the definitions of the macros GO_IF_LEGITIMATE_ADDRESS and ++PRINT_OPERAND_ADDRESS. ++*/ ++#ifdef REG_OK_STRICT ++# define GO_IF_LEGITIMATE_ADDRESS(MODE, X, LABEL) \ ++ do \ ++ { \ ++ if (avr32_legitimate_address(MODE, X, 1)) \ ++ goto LABEL; \ ++ } \ ++ while (0) ++#else ++# define GO_IF_LEGITIMATE_ADDRESS(MODE, X, LABEL) \ ++ do \ ++ { \ ++ if (avr32_legitimate_address(MODE, X, 0)) \ ++ goto LABEL; \ ++ } \ ++ while (0) ++#endif ++ ++ ++ ++/* ++A C compound statement that attempts to replace X with a valid ++memory address for an operand of mode MODE. win will be a ++C statement label elsewhere in the code; the macro definition may use ++ ++ GO_IF_LEGITIMATE_ADDRESS (MODE, X, WIN); ++ ++to avoid further processing if the address has become legitimate. ++ ++X will always be the result of a call to break_out_memory_refs, ++and OLDX will be the operand that was given to that function to produce ++X. ++ ++The code generated by this macro should not alter the substructure of ++X. If it transforms X into a more legitimate form, it ++should assign X (which will always be a C variable) a new value. ++ ++It is not necessary for this macro to come up with a legitimate ++address. The compiler has standard ways of doing so in all cases. In ++fact, it is safe for this macro to do nothing. But often a ++machine-dependent strategy can generate better code. ++*/ ++#define LEGITIMIZE_ADDRESS(X, OLDX, MODE, WIN) \ ++ do \ ++ { \ ++ if (GET_CODE(X) == PLUS \ ++ && GET_CODE(XEXP(X, 0)) == REG \ ++ && GET_CODE(XEXP(X, 1)) == CONST_INT \ ++ && !CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(X, 1)), \ ++ 'K', "Ks16")) \ ++ { \ ++ rtx index = force_reg(SImode, XEXP(X, 1)); \ ++ X = gen_rtx_PLUS( SImode, XEXP(X, 0), index); \ ++ } \ ++ GO_IF_LEGITIMATE_ADDRESS(MODE, X, WIN); \ ++ } \ ++ while(0) ++ ++ ++/* ++A C statement or compound statement with a conditional ++goto LABEL; executed if memory address X (an RTX) can have ++different meanings depending on the machine mode of the memory ++reference it is used for or if the address is valid for some modes ++but not others. ++ ++Autoincrement and autodecrement addresses typically have mode-dependent ++effects because the amount of the increment or decrement is the size ++of the operand being addressed. Some machines have other mode-dependent ++addresses. Many RISC machines have no mode-dependent addresses. ++ ++You may assume that ADDR is a valid address for the machine. ++*/ ++#define GO_IF_MODE_DEPENDENT_ADDRESS(ADDR, LABEL) \ ++ do \ ++ { \ ++ if (GET_CODE (ADDR) == POST_INC \ ++ || GET_CODE (ADDR) == PRE_DEC) \ ++ goto LABEL; \ ++ } \ ++ while (0) ++ ++/* ++A C expression that is nonzero if X is a legitimate constant for ++an immediate operand on the target machine. You can assume that ++X satisfies CONSTANT_P, so you need not check this. In fact, ++'1' is a suitable definition for this macro on machines where ++anything CONSTANT_P is valid. ++*/ ++#define LEGITIMATE_CONSTANT_P(X) avr32_legitimate_constant_p(X) ++ ++ ++/****************************************************************************** ++ * Condition Code Status ++ *****************************************************************************/ ++ ++#define HAVE_conditional_move 1 ++ ++/* ++C code for a data type which is used for declaring the mdep ++component of cc_status. It defaults to int. ++ ++This macro is not used on machines that do not use cc0. ++*/ ++ ++typedef struct ++{ ++ int flags; ++ rtx value; ++ int fpflags; ++ rtx fpvalue; ++} avr32_status_reg; ++ ++ ++#define CC_STATUS_MDEP avr32_status_reg ++ ++/* ++A C expression to initialize the mdep field to "empty". ++The default definition does nothing, since most machines don't use ++the field anyway. If you want to use the field, you should probably ++define this macro to initialize it. ++ ++This macro is not used on machines that do not use cc0. ++*/ ++ ++#define CC_STATUS_MDEP_INIT \ ++ (cc_status.mdep.flags = CC_NONE , cc_status.mdep.value = 0) ++ ++#define FPCC_STATUS_INIT \ ++ (cc_status.mdep.fpflags = CC_NONE , cc_status.mdep.fpvalue = 0) ++ ++/* ++A C compound statement to set the components of cc_status ++appropriately for an insn INSN whose body is EXP. It is ++this macro's responsibility to recognize insns that set the condition ++code as a byproduct of other activity as well as those that explicitly ++set (cc0). ++ ++This macro is not used on machines that do not use cc0. ++ ++If there are insns that do not set the condition code but do alter ++other machine registers, this macro must check to see whether they ++invalidate the expressions that the condition code is recorded as ++reflecting. For example, on the 68000, insns that store in address ++registers do not set the condition code, which means that usually ++NOTICE_UPDATE_CC can leave cc_status unaltered for such ++insns. But suppose that the previous insn set the condition code ++based on location 'a4@@(102)' and the current insn stores a new ++value in 'a4'. Although the condition code is not changed by ++this, it will no longer be true that it reflects the contents of ++'a4@@(102)'. Therefore, NOTICE_UPDATE_CC must alter ++cc_status in this case to say that nothing is known about the ++condition code value. ++ ++The definition of NOTICE_UPDATE_CC must be prepared to deal ++with the results of peephole optimization: insns whose patterns are ++parallel RTXs containing various reg, mem or ++constants which are just the operands. The RTL structure of these ++insns is not sufficient to indicate what the insns actually do. What ++NOTICE_UPDATE_CC should do when it sees one is just to run ++CC_STATUS_INIT. ++ ++A possible definition of NOTICE_UPDATE_CC is to call a function ++that looks at an attribute (see Insn Attributes) named, for example, ++'cc'. This avoids having detailed information about patterns in ++two places, the 'md' file and in NOTICE_UPDATE_CC. ++*/ ++ ++#define NOTICE_UPDATE_CC(EXP, INSN) avr32_notice_update_cc(EXP, INSN) ++ ++ ++ ++ ++/****************************************************************************** ++ * Describing Relative Costs of Operations ++ *****************************************************************************/ ++ ++ ++ ++/* ++A C expression for the cost of moving data of mode MODE from a ++register in class FROM to one in class TO. The classes are ++expressed using the enumeration values such as GENERAL_REGS. A ++value of 2 is the default; other values are interpreted relative to ++that. ++ ++It is not required that the cost always equal 2 when FROM is the ++same as TO; on some machines it is expensive to move between ++registers if they are not general registers. ++ ++If reload sees an insn consisting of a single set between two ++hard registers, and if REGISTER_MOVE_COST applied to their ++classes returns a value of 2, reload does not check to ensure that the ++constraints of the insn are met. Setting a cost of other than 2 will ++allow reload to verify that the constraints are met. You should do this ++if the movm pattern's constraints do not allow such copying. ++*/ ++#define REGISTER_MOVE_COST(MODE, FROM, TO) \ ++ ((GET_MODE_SIZE(MODE) <= 4) ? 2: \ ++ (GET_MODE_SIZE(MODE) <= 8) ? 3: \ ++ 4) ++ ++/* ++A C expression for the cost of moving data of mode MODE between a ++register of class CLASS and memory; IN is zero if the value ++is to be written to memory, nonzero if it is to be read in. This cost ++is relative to those in REGISTER_MOVE_COST. If moving between ++registers and memory is more expensive than between two registers, you ++should define this macro to express the relative cost. ++ ++If you do not define this macro, GCC uses a default cost of 4 plus ++the cost of copying via a secondary reload register, if one is ++needed. If your machine requires a secondary reload register to copy ++between memory and a register of CLASS but the reload mechanism is ++more complex than copying via an intermediate, define this macro to ++reflect the actual cost of the move. ++ ++GCC defines the function memory_move_secondary_cost if ++secondary reloads are needed. It computes the costs due to copying via ++a secondary register. If your machine copies from memory using a ++secondary register in the conventional way but the default base value of ++4 is not correct for your machine, define this macro to add some other ++value to the result of that function. The arguments to that function ++are the same as to this macro. ++*/ ++/* ++ Memory moves are costly ++*/ ++#define MEMORY_MOVE_COST(MODE, CLASS, IN) \ ++ (((IN) ? ((GET_MODE_SIZE(MODE) < 4) ? 4 : \ ++ (GET_MODE_SIZE(MODE) > 8) ? 6 : \ ++ 3) \ ++ : ((GET_MODE_SIZE(MODE) > 8) ? 6 : 3))) ++ ++/* ++A C expression for the cost of a branch instruction. A value of 1 is ++the default; other values are interpreted relative to that. ++*/ ++ /* Try to use conditionals as much as possible */ ++#define BRANCH_COST (TARGET_BRANCH_PRED ? 3 : 4) ++ ++/*A C expression for the maximum number of instructions to execute via conditional ++ execution instructions instead of a branch. A value of BRANCH_COST+1 is the default ++ if the machine does not use cc0, and 1 if it does use cc0.*/ ++#define MAX_CONDITIONAL_EXECUTE 4 ++ ++/* ++Define this macro as a C expression which is nonzero if accessing less ++than a word of memory (i.e.: a char or a short) is no ++faster than accessing a word of memory, i.e., if such access ++require more than one instruction or if there is no difference in cost ++between byte and (aligned) word loads. ++ ++When this macro is not defined, the compiler will access a field by ++finding the smallest containing object; when it is defined, a fullword ++load will be used if alignment permits. Unless bytes accesses are ++faster than word accesses, using word accesses is preferable since it ++may eliminate subsequent memory access if subsequent accesses occur to ++other fields in the same word of the structure, but to different bytes. ++*/ ++#define SLOW_BYTE_ACCESS 1 ++ ++ ++/* ++Define this macro if it is as good or better to call a constant ++function address than to call an address kept in a register. ++*/ ++#define NO_FUNCTION_CSE ++ ++ ++/****************************************************************************** ++ * Adjusting the Instruction Scheduler ++ *****************************************************************************/ ++ ++/***************************************************************************** ++ * Dividing the Output into Sections (Texts, Data, ...) * ++ *****************************************************************************/ ++ ++/* ++A C expression whose value is a string, including spacing, containing the ++assembler operation that should precede instructions and read-only data. ++Normally "\t.text" is right. ++*/ ++#define TEXT_SECTION_ASM_OP "\t.text" ++/* ++A C statement that switches to the default section containing instructions. ++Normally this is not needed, as simply defining TEXT_SECTION_ASM_OP ++is enough. The MIPS port uses this to sort all functions after all data ++declarations. ++*/ ++/* #define TEXT_SECTION */ ++ ++/* ++A C expression whose value is a string, including spacing, containing the ++assembler operation to identify the following data as writable initialized ++data. Normally "\t.data" is right. ++*/ ++#define DATA_SECTION_ASM_OP "\t.data" ++ ++/* ++If defined, a C expression whose value is a string, including spacing, ++containing the assembler operation to identify the following data as ++shared data. If not defined, DATA_SECTION_ASM_OP will be used. ++*/ ++ ++/* ++A C expression whose value is a string, including spacing, containing ++the assembler operation to identify the following data as read-only ++initialized data. ++*/ ++#undef READONLY_DATA_SECTION_ASM_OP ++#define READONLY_DATA_SECTION_ASM_OP \ ++ ((TARGET_USE_RODATA_SECTION) ? \ ++ "\t.section\t.rodata" : \ ++ TEXT_SECTION_ASM_OP ) ++ ++ ++/* ++If defined, a C expression whose value is a string, including spacing, ++containing the assembler operation to identify the following data as ++uninitialized global data. If not defined, and neither ++ASM_OUTPUT_BSS nor ASM_OUTPUT_ALIGNED_BSS are defined, ++uninitialized global data will be output in the data section if ++-fno-common is passed, otherwise ASM_OUTPUT_COMMON will be ++used. ++*/ ++#define BSS_SECTION_ASM_OP "\t.section\t.bss" ++ ++/* ++If defined, a C expression whose value is a string, including spacing, ++containing the assembler operation to identify the following data as ++uninitialized global shared data. If not defined, and ++BSS_SECTION_ASM_OP is, the latter will be used. ++*/ ++/*#define SHARED_BSS_SECTION_ASM_OP "\trseg\tshared_bbs_section:data:noroot(0)\n"*/ ++/* ++If defined, a C expression whose value is a string, including spacing, ++containing the assembler operation to identify the following data as ++initialization code. If not defined, GCC will assume such a section does ++not exist. ++*/ ++#undef INIT_SECTION_ASM_OP ++#define INIT_SECTION_ASM_OP "\t.section\t.init" ++ ++/* ++If defined, a C expression whose value is a string, including spacing, ++containing the assembler operation to identify the following data as ++finalization code. If not defined, GCC will assume such a section does ++not exist. ++*/ ++#undef FINI_SECTION_ASM_OP ++#define FINI_SECTION_ASM_OP "\t.section\t.fini" ++ ++/* ++If defined, an ASM statement that switches to a different section ++via SECTION_OP, calls FUNCTION, and switches back to ++the text section. This is used in crtstuff.c if ++INIT_SECTION_ASM_OP or FINI_SECTION_ASM_OP to calls ++to initialization and finalization functions from the init and fini ++sections. By default, this macro uses a simple function call. Some ++ports need hand-crafted assembly code to avoid dependencies on ++registers initialized in the function prologue or to ensure that ++constant pools don't end up too far way in the text section. ++*/ ++#define CRT_CALL_STATIC_FUNCTION(SECTION_OP, FUNC) \ ++ asm ( SECTION_OP "\n" \ ++ "mcall r6[" USER_LABEL_PREFIX #FUNC "@got]\n" \ ++ TEXT_SECTION_ASM_OP); ++ ++ ++/* ++Define this macro to be an expression with a nonzero value if jump ++tables (for tablejump insns) should be output in the text ++section, along with the assembler instructions. Otherwise, the ++readonly data section is used. ++ ++This macro is irrelevant if there is no separate readonly data section. ++*/ ++/* Put jump tables in text section if we have caches. Otherwise assume that ++ loading data from code memory is slow. */ ++#define JUMP_TABLES_IN_TEXT_SECTION \ ++ (TARGET_CACHES ? 1 : 0) ++ ++ ++/****************************************************************************** ++ * Position Independent Code (PIC) ++ *****************************************************************************/ ++ ++#ifndef AVR32_ALWAYS_PIC ++#define AVR32_ALWAYS_PIC 0 ++#endif ++ ++/* GOT is set to r6 */ ++#define PIC_OFFSET_TABLE_REGNUM INTERNAL_REGNUM(6) ++ ++/* ++A C expression that is nonzero if X is a legitimate immediate ++operand on the target machine when generating position independent code. ++You can assume that X satisfies CONSTANT_P, so you need not ++check this. You can also assume flag_pic is true, so you need not ++check it either. You need not define this macro if all constants ++(including SYMBOL_REF) can be immediate operands when generating ++position independent code. ++*/ ++/* We can't directly access anything that contains a symbol, ++ nor can we indirect via the constant pool. */ ++#define LEGITIMATE_PIC_OPERAND_P(X) avr32_legitimate_pic_operand_p(X) ++ ++ ++/* We need to know when we are making a constant pool; this determines ++ whether data needs to be in the GOT or can be referenced via a GOT ++ offset. */ ++extern int making_const_table; ++ ++/****************************************************************************** ++ * Defining the Output Assembler Language ++ *****************************************************************************/ ++ ++ ++/* ++A C string constant describing how to begin a comment in the target ++assembler language. The compiler assumes that the comment will end at ++the end of the line. ++*/ ++#define ASM_COMMENT_START "# " ++ ++/* ++A C string constant for text to be output before each asm ++statement or group of consecutive ones. Normally this is ++"#APP", which is a comment that has no effect on most ++assemblers but tells the GNU assembler that it must check the lines ++that follow for all valid assembler constructs. ++*/ ++#undef ASM_APP_ON ++#define ASM_APP_ON "#APP\n" ++ ++/* ++A C string constant for text to be output after each asm ++statement or group of consecutive ones. Normally this is ++"#NO_APP", which tells the GNU assembler to resume making the ++time-saving assumptions that are valid for ordinary compiler output. ++*/ ++#undef ASM_APP_OFF ++#define ASM_APP_OFF "#NO_APP\n" ++ ++ ++ ++#define FILE_ASM_OP "\t.file\n" ++#define IDENT_ASM_OP "\t.ident\t" ++#define SET_ASM_OP "\t.set\t" ++ ++ ++/* ++ * Output assembly directives to switch to section name. The section ++ * should have attributes as specified by flags, which is a bit mask ++ * of the SECTION_* flags defined in 'output.h'. If align is nonzero, ++ * it contains an alignment in bytes to be used for the section, ++ * otherwise some target default should be used. Only targets that ++ * must specify an alignment within the section directive need pay ++ * attention to align -- we will still use ASM_OUTPUT_ALIGN. ++ * ++ * NOTE: This one must not be moved to avr32.c ++ */ ++#undef TARGET_ASM_NAMED_SECTION ++#define TARGET_ASM_NAMED_SECTION default_elf_asm_named_section ++ ++ ++/* ++You may define this macro as a C expression. You should define the ++expression to have a nonzero value if GCC should output the constant ++pool for a function before the code for the function, or a zero value if ++GCC should output the constant pool after the function. If you do ++not define this macro, the usual case, GCC will output the constant ++pool before the function. ++*/ ++#define CONSTANT_POOL_BEFORE_FUNCTION 0 ++ ++ ++/* ++Define this macro as a C expression which is nonzero if the constant ++EXP, of type tree, should be output after the code for a ++function. The compiler will normally output all constants before the ++function; you need not define this macro if this is OK. ++*/ ++#define CONSTANT_AFTER_FUNCTION_P(EXP) 1 ++ ++ ++/* ++Define this macro as a C expression which is nonzero if C is ++used as a logical line separator by the assembler. ++ ++If you do not define this macro, the default is that only ++the character ';' is treated as a logical line separator. ++*/ ++#define IS_ASM_LOGICAL_LINE_SEPARATOR(C) ((C) == '\n') ++ ++ ++/** Output of Uninitialized Variables **/ ++ ++/* ++A C statement (sans semicolon) to output to the stdio stream ++STREAM the assembler definition of a common-label named ++NAME whose size is SIZE bytes. The variable ROUNDED ++is the size rounded up to whatever alignment the caller wants. ++ ++Use the expression assemble_name(STREAM, NAME) to ++output the name itself; before and after that, output the additional ++assembler syntax for defining the name, and a newline. ++ ++This macro controls how the assembler definitions of uninitialized ++common global variables are output. ++*/ ++/* ++#define ASM_OUTPUT_COMMON(STREAM, NAME, SIZE, ROUNDED) \ ++ avr32_asm_output_common(STREAM, NAME, SIZE, ROUNDED) ++*/ ++ ++#define ASM_OUTPUT_COMMON(FILE, NAME, SIZE, ROUNDED) \ ++ do \ ++ { \ ++ fputs ("\t.comm ", (FILE)); \ ++ assemble_name ((FILE), (NAME)); \ ++ fprintf ((FILE), ",%d\n", (SIZE)); \ ++ } \ ++ while (0) ++ ++/* ++ * Like ASM_OUTPUT_BSS except takes the required alignment as a ++ * separate, explicit argument. If you define this macro, it is used ++ * in place of ASM_OUTPUT_BSS, and gives you more flexibility in ++ * handling the required alignment of the variable. The alignment is ++ * specified as the number of bits. ++ * ++ * Try to use function asm_output_aligned_bss defined in file varasm.c ++ * when defining this macro. ++ */ ++#define ASM_OUTPUT_ALIGNED_BSS(STREAM, DECL, NAME, SIZE, ALIGNMENT) \ ++ asm_output_aligned_bss (STREAM, DECL, NAME, SIZE, ALIGNMENT) ++ ++/* ++A C statement (sans semicolon) to output to the stdio stream ++STREAM the assembler definition of a local-common-label named ++NAME whose size is SIZE bytes. The variable ROUNDED ++is the size rounded up to whatever alignment the caller wants. ++ ++Use the expression assemble_name(STREAM, NAME) to ++output the name itself; before and after that, output the additional ++assembler syntax for defining the name, and a newline. ++ ++This macro controls how the assembler definitions of uninitialized ++static variables are output. ++*/ ++#define ASM_OUTPUT_LOCAL(FILE, NAME, SIZE, ROUNDED) \ ++ do \ ++ { \ ++ fputs ("\t.lcomm ", (FILE)); \ ++ assemble_name ((FILE), (NAME)); \ ++ fprintf ((FILE), ",%d, %d\n", (SIZE), 2); \ ++ } \ ++ while (0) ++ ++ ++/* ++A C statement (sans semicolon) to output to the stdio stream ++STREAM the assembler definition of a label named NAME. ++Use the expression assemble_name(STREAM, NAME) to ++output the name itself; before and after that, output the additional ++assembler syntax for defining the name, and a newline. ++*/ ++#define ASM_OUTPUT_LABEL(STREAM, NAME) avr32_asm_output_label(STREAM, NAME) ++ ++/* A C string containing the appropriate assembler directive to ++ * specify the size of a symbol, without any arguments. On systems ++ * that use ELF, the default (in 'config/elfos.h') is '"\t.size\t"'; ++ * on other systems, the default is not to define this macro. ++ * ++ * Define this macro only if it is correct to use the default ++ * definitions of ASM_ OUTPUT_SIZE_DIRECTIVE and ++ * ASM_OUTPUT_MEASURED_SIZE for your system. If you need your own ++ * custom definitions of those macros, or if you do not need explicit ++ * symbol sizes at all, do not define this macro. ++ */ ++#define SIZE_ASM_OP "\t.size\t" ++ ++ ++/* ++A C statement (sans semicolon) to output to the stdio stream ++STREAM some commands that will make the label NAME global; ++that is, available for reference from other files. Use the expression ++assemble_name(STREAM, NAME) to output the name ++itself; before and after that, output the additional assembler syntax ++for making that name global, and a newline. ++*/ ++#define GLOBAL_ASM_OP "\t.globl\t" ++ ++ ++ ++/* ++A C expression which evaluates to true if the target supports weak symbols. ++ ++If you don't define this macro, defaults.h provides a default ++definition. If either ASM_WEAKEN_LABEL or ASM_WEAKEN_DECL ++is defined, the default definition is '1'; otherwise, it is ++'0'. Define this macro if you want to control weak symbol support ++with a compiler flag such as -melf. ++*/ ++#define SUPPORTS_WEAK 1 ++ ++/* ++A C statement (sans semicolon) to output to the stdio stream ++STREAM a reference in assembler syntax to a label named ++NAME. This should add '_' to the front of the name, if that ++is customary on your operating system, as it is in most Berkeley Unix ++systems. This macro is used in assemble_name. ++*/ ++#define ASM_OUTPUT_LABELREF(STREAM, NAME) \ ++ avr32_asm_output_labelref(STREAM, NAME) ++ ++ ++ ++/* ++A C expression to assign to OUTVAR (which is a variable of type ++char *) a newly allocated string made from the string ++NAME and the number NUMBER, with some suitable punctuation ++added. Use alloca to get space for the string. ++ ++The string will be used as an argument to ASM_OUTPUT_LABELREF to ++produce an assembler label for an internal static variable whose name is ++NAME. Therefore, the string must be such as to result in valid ++assembler code. The argument NUMBER is different each time this ++macro is executed; it prevents conflicts between similarly-named ++internal static variables in different scopes. ++ ++Ideally this string should not be a valid C identifier, to prevent any ++conflict with the user's own symbols. Most assemblers allow periods ++or percent signs in assembler symbols; putting at least one of these ++between the name and the number will suffice. ++*/ ++#define ASM_FORMAT_PRIVATE_NAME(OUTVAR, NAME, NUMBER) \ ++ do \ ++ { \ ++ (OUTVAR) = (char *) alloca (strlen ((NAME)) + 10); \ ++ sprintf ((OUTVAR), "%s.%d", (NAME), (NUMBER)); \ ++ } \ ++ while (0) ++ ++ ++/** Macros Controlling Initialization Routines **/ ++ ++ ++/* ++If defined, main will not call __main as described above. ++This macro should be defined for systems that control start-up code ++on a symbol-by-symbol basis, such as OSF/1, and should not ++be defined explicitly for systems that support INIT_SECTION_ASM_OP. ++*/ ++/* ++ __main is not defined when debugging. ++*/ ++#define HAS_INIT_SECTION ++ ++ ++/** Output of Assembler Instructions **/ ++ ++/* ++A C initializer containing the assembler's names for the machine ++registers, each one as a C string constant. This is what translates ++register numbers in the compiler into assembler language. ++*/ ++ ++#define REGISTER_NAMES \ ++{ \ ++ "pc", "lr", \ ++ "sp", "r12", \ ++ "r11", "r10", \ ++ "r9", "r8", \ ++ "r7", "r6", \ ++ "r5", "r4", \ ++ "r3", "r2", \ ++ "r1", "r0", \ ++ "f15","f14", \ ++ "f13","f12", \ ++ "f11","f10", \ ++ "f9", "f8", \ ++ "f7", "f6", \ ++ "f5", "f4", \ ++ "f3", "f2", \ ++ "f1", "f0" \ ++} ++ ++/* ++A C compound statement to output to stdio stream STREAM the ++assembler syntax for an instruction operand X. X is an ++RTL expression. ++ ++CODE is a value that can be used to specify one of several ways ++of printing the operand. It is used when identical operands must be ++printed differently depending on the context. CODE comes from ++the '%' specification that was used to request printing of the ++operand. If the specification was just '%digit' then ++CODE is 0; if the specification was '%ltr digit' ++then CODE is the ASCII code for ltr. ++ ++If X is a register, this macro should print the register's name. ++The names can be found in an array reg_names whose type is ++char *[]. reg_names is initialized from REGISTER_NAMES. ++ ++When the machine description has a specification '%punct' ++(a '%' followed by a punctuation character), this macro is called ++with a null pointer for X and the punctuation character for ++CODE. ++*/ ++#define PRINT_OPERAND(STREAM, X, CODE) avr32_print_operand(STREAM, X, CODE) ++ ++/* A C statement to be executed just prior to the output of ++ assembler code for INSN, to modify the extracted operands so ++ they will be output differently. ++ ++ Here the argument OPVEC is the vector containing the operands ++ extracted from INSN, and NOPERANDS is the number of elements of ++ the vector which contain meaningful data for this insn. ++ The contents of this vector are what will be used to convert the insn ++ template into assembler code, so you can change the assembler output ++ by changing the contents of the vector. */ ++#define FINAL_PRESCAN_INSN(INSN, OPVEC, NOPERANDS) \ ++ avr32_final_prescan_insn ((INSN), (OPVEC), (NOPERANDS)) ++ ++/* ++A C expression which evaluates to true if CODE is a valid ++punctuation character for use in the PRINT_OPERAND macro. If ++PRINT_OPERAND_PUNCT_VALID_P is not defined, it means that no ++punctuation characters (except for the standard one, '%') are used ++in this way. ++*/ ++/* ++ 'm' refers to the most significant word in a two-register mode. ++*/ ++#define PRINT_OPERAND_PUNCT_VALID_P(CODE) ((CODE) == 'm' || (CODE) == 'e') ++ ++/* ++A C compound statement to output to stdio stream STREAM the ++assembler syntax for an instruction operand that is a memory reference ++whose address is X. X is an RTL expression. ++ ++On some machines, the syntax for a symbolic address depends on the ++section that the address refers to. On these machines, define the macro ++ENCODE_SECTION_INFO to store the information into the ++symbol_ref, and then check for it here. (see Assembler Format.) ++*/ ++#define PRINT_OPERAND_ADDRESS(STREAM, X) avr32_print_operand_address(STREAM, X) ++ ++ ++/** Output of Dispatch Tables **/ ++ ++/* ++ * A C statement to output to the stdio stream stream an assembler ++ * pseudo-instruction to generate a difference between two ++ * labels. value and rel are the numbers of two internal labels. The ++ * definitions of these labels are output using ++ * (*targetm.asm_out.internal_label), and they must be printed in the ++ * same way here. For example, ++ * ++ * fprintf (stream, "\t.word L%d-L%d\n", ++ * value, rel) ++ * ++ * You must provide this macro on machines where the addresses in a ++ * dispatch table are relative to the table's own address. If defined, ++ * GCC will also use this macro on all machines when producing ++ * PIC. body is the body of the ADDR_DIFF_VEC; it is provided so that ++ * the mode and flags can be read. ++ */ ++#define ASM_OUTPUT_ADDR_DIFF_ELT(STREAM, BODY, VALUE, REL) \ ++ fprintf(STREAM, "\tbral\t%sL%d\n", LOCAL_LABEL_PREFIX, VALUE) ++ ++/* ++This macro should be provided on machines where the addresses ++in a dispatch table are absolute. ++ ++The definition should be a C statement to output to the stdio stream ++STREAM an assembler pseudo-instruction to generate a reference to ++a label. VALUE is the number of an internal label whose ++definition is output using ASM_OUTPUT_INTERNAL_LABEL. ++For example, ++ ++fprintf(STREAM, "\t.word L%d\n", VALUE) ++*/ ++ ++#define ASM_OUTPUT_ADDR_VEC_ELT(STREAM, VALUE) \ ++ fprintf(STREAM, "\t.long %sL%d\n", LOCAL_LABEL_PREFIX, VALUE) ++ ++/** Assembler Commands for Exception Regions */ ++ ++/* ToDo: All of this subsection */ ++ ++/** Assembler Commands for Alignment */ ++ ++ ++/* ++A C statement to output to the stdio stream STREAM an assembler ++command to advance the location counter to a multiple of 2 to the ++POWER bytes. POWER will be a C expression of type int. ++*/ ++#define ASM_OUTPUT_ALIGN(STREAM, POWER) \ ++ do \ ++ { \ ++ if ((POWER) != 0) \ ++ fprintf(STREAM, "\t.align\t%d\n", POWER); \ ++ } \ ++ while (0) ++ ++/* ++Like ASM_OUTPUT_ALIGN, except that the \nop" instruction is used for padding, if ++necessary. ++*/ ++#define ASM_OUTPUT_ALIGN_WITH_NOP(STREAM, POWER) \ ++ fprintf(STREAM, "\t.balignw\t%d, 0xd703\n", (1 << POWER)) ++ ++ ++ ++/****************************************************************************** ++ * Controlling Debugging Information Format ++ *****************************************************************************/ ++ ++/* How to renumber registers for dbx and gdb. */ ++#define DBX_REGISTER_NUMBER(REGNO) ASM_REGNUM (REGNO) ++ ++/* The DWARF 2 CFA column which tracks the return address. */ ++#define DWARF_FRAME_RETURN_COLUMN DWARF_FRAME_REGNUM(LR_REGNUM) ++ ++/* ++Define this macro if GCC should produce dwarf version 2 format ++debugging output in response to the -g option. ++ ++To support optional call frame debugging information, you must also ++define INCOMING_RETURN_ADDR_RTX and either set ++RTX_FRAME_RELATED_P on the prologue insns if you use RTL for the ++prologue, or call dwarf2out_def_cfa and dwarf2out_reg_save ++as appropriate from TARGET_ASM_FUNCTION_PROLOGUE if you don't. ++*/ ++#define DWARF2_DEBUGGING_INFO 1 ++ ++ ++#define DWARF2_ASM_LINE_DEBUG_INFO 1 ++#define DWARF2_FRAME_INFO 1 ++ ++ ++/****************************************************************************** ++ * Miscellaneous Parameters ++ *****************************************************************************/ ++ ++/* ToDo: a lot */ ++ ++/* ++An alias for a machine mode name. This is the machine mode that ++elements of a jump-table should have. ++*/ ++#define CASE_VECTOR_MODE SImode ++ ++/* ++Define this macro to be a C expression to indicate when jump-tables ++should contain relative addresses. If jump-tables never contain ++relative addresses, then you need not define this macro. ++*/ ++#define CASE_VECTOR_PC_RELATIVE 0 ++ ++/* Increase the threshold for using table jumps on the UC arch. */ ++#define CASE_VALUES_THRESHOLD ((avr32_arch->arch_type == ARCH_TYPE_AVR32_UC) ? 7 : 4) ++ ++/* ++The maximum number of bytes that a single instruction can move quickly ++between memory and registers or between two memory locations. ++*/ ++#define MOVE_MAX (2*UNITS_PER_WORD) ++ ++ ++/* A C expression that is nonzero if on this machine the number of bits actually used ++ for the count of a shift operation is equal to the number of bits needed to represent ++ the size of the object being shifted. When this macro is nonzero, the compiler will ++ assume that it is safe to omit a sign-extend, zero-extend, and certain bitwise 'and' ++ instructions that truncates the count of a shift operation. On machines that have ++ instructions that act on bit-fields at variable positions, which may include 'bit test' ++ 378 GNU Compiler Collection (GCC) Internals ++ instructions, a nonzero SHIFT_COUNT_TRUNCATED also enables deletion of truncations ++ of the values that serve as arguments to bit-field instructions. ++ If both types of instructions truncate the count (for shifts) and position (for bit-field ++ operations), or if no variable-position bit-field instructions exist, you should define ++ this macro. ++ However, on some machines, such as the 80386 and the 680x0, truncation only applies ++ to shift operations and not the (real or pretended) bit-field operations. Define SHIFT_ ++ COUNT_TRUNCATED to be zero on such machines. Instead, add patterns to the 'md' file ++ that include the implied truncation of the shift instructions. ++ You need not de ne this macro if it would always have the value of zero. */ ++#define SHIFT_COUNT_TRUNCATED 1 ++ ++/* ++A C expression which is nonzero if on this machine it is safe to ++convert an integer of INPREC bits to one of OUTPREC ++bits (where OUTPREC is smaller than INPREC) by merely ++operating on it as if it had only OUTPREC bits. ++ ++On many machines, this expression can be 1. ++ ++When TRULY_NOOP_TRUNCATION returns 1 for a pair of sizes for ++modes for which MODES_TIEABLE_P is 0, suboptimal code can result. ++If this is the case, making TRULY_NOOP_TRUNCATION return 0 in ++such cases may improve things. ++*/ ++#define TRULY_NOOP_TRUNCATION(OUTPREC, INPREC) 1 ++ ++/* ++An alias for the machine mode for pointers. On most machines, define ++this to be the integer mode corresponding to the width of a hardware ++pointer; SImode on 32-bit machine or DImode on 64-bit machines. ++On some machines you must define this to be one of the partial integer ++modes, such as PSImode. ++ ++The width of Pmode must be at least as large as the value of ++POINTER_SIZE. If it is not equal, you must define the macro ++POINTERS_EXTEND_UNSIGNED to specify how pointers are extended ++to Pmode. ++*/ ++#define Pmode SImode ++ ++/* ++An alias for the machine mode used for memory references to functions ++being called, in call RTL expressions. On most machines this ++should be QImode. ++*/ ++#define FUNCTION_MODE SImode ++ ++ ++#define REG_S_P(x) \ ++ (REG_P (x) || (GET_CODE (x) == SUBREG && REG_P (XEXP (x, 0)))) ++ ++ ++/* If defined, modifies the length assigned to instruction INSN as a ++ function of the context in which it is used. LENGTH is an lvalue ++ that contains the initially computed length of the insn and should ++ be updated with the correct length of the insn. */ ++#define ADJUST_INSN_LENGTH(INSN, LENGTH) \ ++ ((LENGTH) = avr32_adjust_insn_length ((INSN), (LENGTH))) ++ ++ ++#define CLZ_DEFINED_VALUE_AT_ZERO(mode, value) \ ++ (value = 32, (mode == SImode)) ++ ++#define CTZ_DEFINED_VALUE_AT_ZERO(mode, value) \ ++ (value = 32, (mode == SImode)) ++ ++#define UNITS_PER_SIMD_WORD UNITS_PER_WORD ++ ++#define STORE_FLAG_VALUE 1 ++ ++enum avr32_builtins ++{ ++ AVR32_BUILTIN_MTSR, ++ AVR32_BUILTIN_MFSR, ++ AVR32_BUILTIN_MTDR, ++ AVR32_BUILTIN_MFDR, ++ AVR32_BUILTIN_CACHE, ++ AVR32_BUILTIN_SYNC, ++ AVR32_BUILTIN_TLBR, ++ AVR32_BUILTIN_TLBS, ++ AVR32_BUILTIN_TLBW, ++ AVR32_BUILTIN_BREAKPOINT, ++ AVR32_BUILTIN_XCHG, ++ AVR32_BUILTIN_LDXI, ++ AVR32_BUILTIN_BSWAP16, ++ AVR32_BUILTIN_BSWAP32, ++ AVR32_BUILTIN_COP, ++ AVR32_BUILTIN_MVCR_W, ++ AVR32_BUILTIN_MVRC_W, ++ AVR32_BUILTIN_MVCR_D, ++ AVR32_BUILTIN_MVRC_D, ++ AVR32_BUILTIN_MULSATHH_H, ++ AVR32_BUILTIN_MULSATHH_W, ++ AVR32_BUILTIN_MULSATRNDHH_H, ++ AVR32_BUILTIN_MULSATRNDWH_W, ++ AVR32_BUILTIN_MULSATWH_W, ++ AVR32_BUILTIN_MACSATHH_W, ++ AVR32_BUILTIN_SATADD_H, ++ AVR32_BUILTIN_SATSUB_H, ++ AVR32_BUILTIN_SATADD_W, ++ AVR32_BUILTIN_SATSUB_W, ++ AVR32_BUILTIN_MULWH_D, ++ AVR32_BUILTIN_MULNWH_D, ++ AVR32_BUILTIN_MACWH_D, ++ AVR32_BUILTIN_MACHH_D, ++ AVR32_BUILTIN_MUSFR, ++ AVR32_BUILTIN_MUSTR, ++ AVR32_BUILTIN_SATS, ++ AVR32_BUILTIN_SATU, ++ AVR32_BUILTIN_SATRNDS, ++ AVR32_BUILTIN_SATRNDU ++}; ++ ++ ++#define FLOAT_LIB_COMPARE_RETURNS_BOOL(MODE, COMPARISON) \ ++ ((MODE == SFmode) || (MODE == DFmode)) ++ ++#define RENAME_LIBRARY_SET ".set" ++ ++/* Make ABI_NAME an alias for __GCC_NAME. */ ++#define RENAME_LIBRARY(GCC_NAME, ABI_NAME) \ ++ __asm__ (".globl\t__avr32_" #ABI_NAME "\n" \ ++ ".set\t__avr32_" #ABI_NAME \ ++ ", __" #GCC_NAME "\n"); ++ ++/* Give libgcc functions avr32 ABI name. */ ++#ifdef L_muldi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (muldi3, mul64) ++#endif ++#ifdef L_divdi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (divdi3, sdiv64) ++#endif ++#ifdef L_udivdi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (udivdi3, udiv64) ++#endif ++#ifdef L_moddi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (moddi3, smod64) ++#endif ++#ifdef L_umoddi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (umoddi3, umod64) ++#endif ++#ifdef L_ashldi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (ashldi3, lsl64) ++#endif ++#ifdef L_lshrdi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (lshrdi3, lsr64) ++#endif ++#ifdef L_ashrdi3 ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (ashrdi3, asr64) ++#endif ++ ++#ifdef L_fixsfdi ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (fixsfdi, f32_to_s64) ++#endif ++#ifdef L_fixunssfdi ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (fixunssfdi, f32_to_u64) ++#endif ++#ifdef L_floatdidf ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (floatdidf, s64_to_f64) ++#endif ++#ifdef L_floatdisf ++#define DECLARE_LIBRARY_RENAMES RENAME_LIBRARY (floatdisf, s64_to_f32) ++#endif ++ ++#endif +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.md +--- gcc-4.2.1/gcc/config/avr32/avr32.md 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.md 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,4606 @@ ++;; AVR32 machine description file. ++;; Copyright 2003-2006 Atmel Corporation. ++;; ++;; Written by Ronny Pedersen, Atmel Norway, ++;; ++;; This file is part of GCC. ++;; ++;; This program is free software; you can redistribute it and/or modify ++;; it under the terms of the GNU General Public License as published by ++;; the Free Software Foundation; either version 2 of the License, or ++;; (at your option) any later version. ++;; ++;; This program is distributed in the hope that it will be useful, ++;; but WITHOUT ANY WARRANTY; without even the implied warranty of ++;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++;; GNU General Public License for more details. ++;; ++;; You should have received a copy of the GNU General Public License ++;; along with this program; if not, write to the Free Software ++;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. ++ ++;; -*- Mode: Scheme -*- ++ ++(define_attr "type" "alu,alu2,alu_sat,mulhh,mulwh,mulww_w,mulww_d,div,machh_w,macww_w,macww_d,branch,call,load,load_rm,store,load2,load4,store2,store4,fmul,fcmps,fcmpd,fcast,fmv,fmvcpu,fldd,fstd,flds,fsts,fstm" ++ (const_string "alu")) ++ ++ ++(define_attr "cc" "none,set_vncz,set_ncz,set_cz,set_z,bld,compare,cmp_cond_insn,clobber,call_set,fpcompare,from_fpcc" ++ (const_string "none")) ++ ++ ++; NB! Keep this in sync with enum architecture_type in avr32.h ++(define_attr "pipeline" "ap,uc" ++ (const (symbol_ref "avr32_arch->arch_type"))) ++ ++(define_attr "length" "" ++ (const_int 4)) ++ ++ ++;; Uses of UNSPEC in this file: ++(define_constants ++ [(UNSPEC_PUSHM 0) ++ (UNSPEC_POPM 1) ++ (UNSPEC_UDIVMODSI4_INTERNAL 2) ++ (UNSPEC_DIVMODSI4_INTERNAL 3) ++ (UNSPEC_STM 4) ++ (UNSPEC_LDM 5) ++ (UNSPEC_MOVSICC 6) ++ (UNSPEC_ADDSICC 7) ++ (UNSPEC_COND_MI 8) ++ (UNSPEC_COND_PL 9) ++ (UNSPEC_PIC_SYM 10) ++ (UNSPEC_PIC_BASE 11) ++ (UNSPEC_STORE_MULTIPLE 12) ++ (UNSPEC_STMFP 13) ++ (UNSPEC_FPCC_TO_REG 14) ++ (UNSPEC_REG_TO_CC 15) ++ (UNSPEC_FORCE_MINIPOOL 16) ++ (UNSPEC_SATS 17) ++ (UNSPEC_SATU 18) ++ (UNSPEC_SATRNDS 19) ++ (UNSPEC_SATRNDU 20) ++ ]) ++ ++(define_constants ++ [(VUNSPEC_EPILOGUE 0) ++ (VUNSPEC_CACHE 1) ++ (VUNSPEC_MTSR 2) ++ (VUNSPEC_MFSR 3) ++ (VUNSPEC_BLOCKAGE 4) ++ (VUNSPEC_SYNC 5) ++ (VUNSPEC_TLBR 6) ++ (VUNSPEC_TLBW 7) ++ (VUNSPEC_TLBS 8) ++ (VUNSPEC_BREAKPOINT 9) ++ (VUNSPEC_MTDR 10) ++ (VUNSPEC_MFDR 11) ++ (VUNSPEC_MVCR 12) ++ (VUNSPEC_MVRC 13) ++ (VUNSPEC_COP 14) ++ (VUNSPEC_ALIGN 15) ++ (VUNSPEC_POOL_START 16) ++ (VUNSPEC_POOL_END 17) ++ (VUNSPEC_POOL_4 18) ++ (VUNSPEC_POOL_8 19) ++ (VUNSPEC_POOL_16 20) ++ (VUNSPEC_MUSFR 21) ++ (VUNSPEC_MUSTR 22) ++ (VUNSPEC_SYNC_CMPXCHG 23) ++ (VUNSPEC_SYNC_SET_LOCK_AND_LOAD 24) ++ (VUNSPEC_SYNC_STORE_IF_LOCK 25) ++ (VUNSPEC_EH_RETURN 26) ++ (VUNSPEC_FRS 27) ++ ]) ++ ++(define_constants ++ [ ++ ;; R7 = 15-7 = 8 ++ (FP_REGNUM 8) ++ ;; Return Register = R12 = 15 - 12 = 3 ++ (RETVAL_REGNUM 3) ++ ;; SP = R13 = 15 - 13 = 2 ++ (SP_REGNUM 2) ++ ;; LR = R14 = 15 - 14 = 1 ++ (LR_REGNUM 1) ++ ;; PC = R15 = 15 - 15 = 0 ++ (PC_REGNUM 0) ++ ;; FPSR = GENERAL_REGS + 1 = 17 ++ (FPCC_REGNUM 17) ++ ]) ++ ++ ++ ++ ++;;****************************************************************************** ++;; Macros ++;;****************************************************************************** ++ ++;; Integer Modes for basic alu insns ++(define_mode_macro INTM [SI HI QI]) ++(define_mode_attr alu_cc_attr [(SI "set_vncz") (HI "clobber") (QI "clobber")]) ++ ++;; Move word modes ++(define_mode_macro MOVM [SI V2HI V4QI]) ++ ++;; For mov/addcc insns ++(define_mode_macro ADDCC [SI HI QI]) ++(define_mode_macro MOVCC [SI HI QI]) ++(define_mode_macro CMP [DI SI HI QI]) ++(define_mode_attr cmp_constraint [(DI "rKu20") (SI "rKs21") (HI "r") (QI "r")]) ++(define_mode_attr cmp_predicate [(DI "register_immediate_operand") ++ (SI "register_immediate_operand") ++ (HI "register_operand") ++ (QI "register_operand")]) ++ ++;; For all conditional insns ++(define_code_macro any_cond [eq ne gt ge lt le gtu geu ltu leu]) ++(define_code_attr cond [(eq "eq") (ne "ne") (gt "gt") (ge "ge") (lt "lt") (le "le") ++ (gtu "hi") (geu "hs") (ltu "lo") (leu "ls")]) ++(define_code_attr invcond [(eq "ne") (ne "eq") (gt "le") (ge "lt") (lt "ge") (le "gt") ++ (gtu "ls") (geu "lo") (ltu "hs") (leu "hi")]) ++ ++;; For logical operations ++(define_code_macro logical [and ior xor]) ++(define_code_attr logical_insn [(and "and") (ior "or") (xor "eor")]) ++ ++;; Load the predicates ++(include "predicates.md") ++ ++ ++;;****************************************************************************** ++;; Automaton pipeline description for avr32 ++;;****************************************************************************** ++ ++(define_automaton "avr32_ap") ++ ++ ++(define_cpu_unit "is" "avr32_ap") ++(define_cpu_unit "a1,m1,da" "avr32_ap") ++(define_cpu_unit "a2,m2,d" "avr32_ap") ++ ++;;Alu instructions ++(define_insn_reservation "alu_op" 1 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "alu")) ++ "is,a1,a2") ++ ++(define_insn_reservation "alu2_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "alu2")) ++ "is,is+a1,a1+a2,a2") ++ ++(define_insn_reservation "alu_sat_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "alu_sat")) ++ "is,a1,a2") ++ ++ ++;;Mul instructions ++(define_insn_reservation "mulhh_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "mulhh,mulwh")) ++ "is,m1,m2") ++ ++(define_insn_reservation "mulww_w_op" 3 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "mulww_w")) ++ "is,m1,m1+m2,m2") ++ ++(define_insn_reservation "mulww_d_op" 5 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "mulww_d")) ++ "is,m1,m1+m2,m1+m2,m2,m2") ++ ++(define_insn_reservation "div_op" 33 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "div")) ++ "is,m1,m1*31 + m2*31,m2") ++ ++(define_insn_reservation "machh_w_op" 3 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "machh_w")) ++ "is*2,m1,m2") ++ ++ ++(define_insn_reservation "macww_w_op" 4 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "macww_w")) ++ "is*2,m1,m1,m2") ++ ++ ++(define_insn_reservation "macww_d_op" 6 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "macww_d")) ++ "is*2,m1,m1+m2,m1+m2,m2") ++ ++;;Bypasses for Mac instructions, because of accumulator cache. ++;;Set latency as low as possible in order to let the compiler let ++;;mul -> mac and mac -> mac combinations which use the same ++;;accumulator cache be placed close together to avoid any ++;;instructions which can ruin the accumulator cache come inbetween. ++(define_bypass 4 "machh_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++(define_bypass 5 "macww_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++(define_bypass 7 "macww_d_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++ ++(define_bypass 3 "mulhh_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++(define_bypass 4 "mulww_w_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++(define_bypass 6 "mulww_d_op" "alu_op,alu2_op,alu_sat_op,load_op" "avr32_mul_waw_bypass") ++ ++ ++;;Bypasses for all mul/mac instructions followed by an instruction ++;;which reads the output AND writes the result to the same register. ++;;This will generate an Write After Write hazard which gives an ++;;extra cycle before the result is ready. ++(define_bypass 0 "machh_w_op" "machh_w_op" "avr32_valid_macmac_bypass") ++(define_bypass 0 "macww_w_op" "macww_w_op" "avr32_valid_macmac_bypass") ++(define_bypass 0 "macww_d_op" "macww_d_op" "avr32_valid_macmac_bypass") ++ ++(define_bypass 0 "mulhh_op" "machh_w_op" "avr32_valid_mulmac_bypass") ++(define_bypass 0 "mulww_w_op" "macww_w_op" "avr32_valid_mulmac_bypass") ++(define_bypass 0 "mulww_d_op" "macww_d_op" "avr32_valid_mulmac_bypass") ++ ++;;Branch and call instructions ++;;We assume that all branches and rcalls are predicted correctly :-) ++;;while calls use a lot of cycles. ++(define_insn_reservation "branch_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "branch")) ++ "nothing") ++ ++(define_insn_reservation "call_op" 10 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "call")) ++ "nothing") ++ ++ ++;;Load store instructions ++(define_insn_reservation "load_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "load")) ++ "is,da,d") ++ ++(define_insn_reservation "load_rm_op" 3 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "load_rm")) ++ "is,da,d") ++ ++ ++(define_insn_reservation "store_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "store")) ++ "is,da,d") ++ ++ ++(define_insn_reservation "load_double_op" 3 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "load2")) ++ "is,da,da+d,d") ++ ++(define_insn_reservation "load_quad_op" 4 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "load4")) ++ "is,da,da+d,da+d,d") ++ ++(define_insn_reservation "store_double_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "store2")) ++ "is,da,da+d,d") ++ ++ ++(define_insn_reservation "store_quad_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "store4")) ++ "is,da,da+d,da+d,d") ++ ++;;For store the operand to write to memory is read in d and ++;;the real latency between any instruction and a store is therefore ++;;one less than for the instructions which reads the operands in the first ++;;excecution stage ++(define_bypass 2 "load_double_op" "store_double_op" "avr32_store_bypass") ++(define_bypass 3 "load_quad_op" "store_quad_op" "avr32_store_bypass") ++(define_bypass 1 "load_op" "store_op" "avr32_store_bypass") ++(define_bypass 2 "load_rm_op" "store_op" "avr32_store_bypass") ++(define_bypass 1 "alu_sat_op" "store_op" "avr32_store_bypass") ++(define_bypass 1 "alu2_op" "store_op" "avr32_store_bypass") ++(define_bypass 1 "mulhh_op" "store_op" "avr32_store_bypass") ++(define_bypass 2 "mulww_w_op" "store_op" "avr32_store_bypass") ++(define_bypass 4 "mulww_d_op" "store_op" "avr32_store_bypass" ) ++(define_bypass 2 "machh_w_op" "store_op" "avr32_store_bypass") ++(define_bypass 3 "macww_w_op" "store_op" "avr32_store_bypass") ++(define_bypass 5 "macww_d_op" "store_op" "avr32_store_bypass") ++ ++ ++; Bypass for load double operation. If only the first loaded word is needed ++; then the latency is 2 ++(define_bypass 2 "load_double_op" ++ "load_op,load_rm_op,alu_sat_op, alu2_op, alu_op, mulhh_op, mulww_w_op, ++ mulww_d_op, machh_w_op, macww_w_op, macww_d_op" ++ "avr32_valid_load_double_bypass") ++ ++; Bypass for load quad operation. If only the first or second loaded word is needed ++; we set the latency to 2 ++(define_bypass 2 "load_quad_op" ++ "load_op,load_rm_op,alu_sat_op, alu2_op, alu_op, mulhh_op, mulww_w_op, ++ mulww_d_op, machh_w_op, macww_w_op, macww_d_op" ++ "avr32_valid_load_quad_bypass") ++ ++ ++;;****************************************************************************** ++;; End of Automaton pipeline description for avr32 ++;;****************************************************************************** ++ ++ ++ ++;;============================================================================= ++;; move ++;;----------------------------------------------------------------------------- ++ ++;;== char - 8 bits ============================================================ ++(define_expand "movqi" ++ [(set (match_operand:QI 0 "nonimmediate_operand" "") ++ (match_operand:QI 1 "general_operand" ""))] ++ "" ++ { ++ if ( !no_new_pseudos ){ ++ if (GET_CODE (operands[1]) == MEM && optimize){ ++ rtx reg = gen_reg_rtx (SImode); ++ ++ emit_insn (gen_zero_extendqisi2 (reg, operands[1])); ++ operands[1] = gen_lowpart (QImode, reg); ++ } ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) == MEM) ++ operands[1] = force_reg (QImode, operands[1]); ++ } ++ ++ }) ++ ++(define_insn "*movqi_internal" ++ [(set (match_operand:QI 0 "nonimmediate_operand" "=r,r,m,r") ++ (match_operand:QI 1 "general_operand" "rKs08,m,r,i"))] ++ "" ++ "@ ++ mov\t%0, %1 ++ ld.ub\t%0, %1 ++ st.b\t%0, %1 ++ mov\t%0, %1" ++ [(set_attr "length" "2,4,4,4") ++ (set_attr "type" "alu,load_rm,store,alu")]) ++ ++ ++ ++;;== short - 16 bits ========================================================== ++(define_expand "movhi" ++ [(set (match_operand:HI 0 "nonimmediate_operand" "") ++ (match_operand:HI 1 "general_operand" ""))] ++ "" ++ { ++ if ( !no_new_pseudos ){ ++ if (GET_CODE (operands[1]) == MEM && optimize){ ++ rtx reg = gen_reg_rtx (SImode); ++ ++ emit_insn (gen_extendhisi2 (reg, operands[1])); ++ operands[1] = gen_lowpart (HImode, reg); ++ } ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) == MEM) ++ operands[1] = force_reg (HImode, operands[1]); ++ } ++ ++ }) ++ ++(define_insn "*movhi_internal" ++ [(set (match_operand:HI 0 "nonimmediate_operand" "=r,r,m,r") ++ (match_operand:HI 1 "general_operand" "rKs08,m,r,i"))] ++ "" ++ "@ ++ mov\t%0, %1 ++ ld.sh\t%0, %1 ++ st.h\t%0, %1 ++ mov\t%0, %1" ++ [(set_attr "length" "2,4,4,4") ++ (set_attr "type" "alu,load_rm,store,alu")]) ++ ++ ++;;== int - 32 bits ============================================================ ++ ++(define_expand "movmisalignsi" ++ [(set (match_operand:SI 0 "nonimmediate_operand" "") ++ (match_operand:SI 1 "nonimmediate_operand" ""))] ++ "TARGET_UNALIGNED_WORD" ++ { ++ } ++) ++ ++(define_expand "mov" ++ [(set (match_operand:MOVM 0 "register_operand" "") ++ (match_operand:MOVM 1 "general_operand" ""))] ++ "" ++ { ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) == MEM) ++ operands[1] = force_reg (mode, operands[1]); ++ ++ ++ /* Check for out of range immediate constants as these may ++ occur during reloading, since it seems like reload does ++ not check if the immediate is legitimate. Don't know if ++ this is a bug? */ ++ if ( reload_in_progress ++ && avr32_arch->arch_type != ARCH_TYPE_AVR32_UC ++ && GET_CODE(operands[1]) == CONST_INT ++ && !avr32_const_ok_for_constraint_p(INTVAL(operands[1]), 'K', "Ks21") ){ ++ operands[1] = force_const_mem(SImode, operands[1]); ++ } ++ ++ if ( (flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS) ++ && !avr32_legitimate_pic_operand_p(operands[1]) ) ++ operands[1] = legitimize_pic_address (operands[1], mode, ++ (no_new_pseudos ? operands[0] : 0)); ++ else if ( flag_pic && avr32_address_operand(operands[1], GET_MODE(operands[1])) ) ++ /* If we have an address operand then this function uses the pic register. */ ++ current_function_uses_pic_offset_table = 1; ++ }) ++ ++ ++(define_insn "mov_internal" ++ [(set (match_operand:MOVM 0 "nonimmediate_operand" "=r, r, r,r,m,r") ++ (match_operand:MOVM 1 "general_operand" "rKs08,Ks21,n,m,r,W"))] ++ "" ++ { ++ switch (which_alternative) { ++ case 0: ++ case 1: return "mov\t%0, %1"; ++ case 2: return "mov\t%0, lo(%1)\;orh\t%0,hi(%1)"; ++ case 3: ++ if ( (REG_P(XEXP(operands[1], 0)) ++ && REGNO(XEXP(operands[1], 0)) == SP_REGNUM) ++ || (GET_CODE(XEXP(operands[1], 0)) == PLUS ++ && REGNO(XEXP(XEXP(operands[1], 0), 0)) == SP_REGNUM ++ && GET_CODE(XEXP(XEXP(operands[1], 0), 1)) == CONST_INT ++ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) % 4 == 0 ++ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) <= 0x1FC) ) ++ return "lddsp\t%0, %1"; ++ else if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1])) ) ++ return "lddpc\t%0, %1"; ++ else ++ return "ld.w\t%0, %1"; ++ case 4: ++ if ( (REG_P(XEXP(operands[0], 0)) ++ && REGNO(XEXP(operands[0], 0)) == SP_REGNUM) ++ || (GET_CODE(XEXP(operands[0], 0)) == PLUS ++ && REGNO(XEXP(XEXP(operands[0], 0), 0)) == SP_REGNUM ++ && GET_CODE(XEXP(XEXP(operands[0], 0), 1)) == CONST_INT ++ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) % 4 == 0 ++ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) <= 0x1FC) ) ++ return "stdsp\t%0, %1"; ++ else ++ return "st.w\t%0, %1"; ++ case 5: ++ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) ++ return "lda.w\t%0, %1"; ++ else ++ return "ld.w\t%0, r6[%1@got]"; ++ default: ++ abort(); ++ } ++ } ++ ++ [(set_attr "length" "2,4,8,4,4,8") ++ (set_attr "type" "alu,alu,alu2,load,store,load") ++ (set_attr "cc" "none,none,set_z,none,none,clobber")]) ++ ++ ++;; These instructions are for loading constants which cannot be loaded ++;; directly from the constant pool because the offset is too large ++;; high and lo_sum are used even tough for our case it should be ++;; low and high sum :-) ++(define_insn "mov_symbol_lo" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (high:SI (match_operand:SI 1 "immediate_operand" "i" )))] ++ "" ++ "mov\t%0, lo(%1)" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4")] ++) ++ ++(define_insn "add_symbol_hi" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (lo_sum:SI (match_dup 0) ++ (match_operand:SI 1 "immediate_operand" "i" )))] ++ "" ++ "orh\t%0, hi(%1)" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4")] ++) ++ ++ ++ ++;; When generating pic, we need to load the symbol offset into a register. ++;; So that the optimizer does not confuse this with a normal symbol load ++;; we use an unspec. The offset will be loaded from a constant pool entry, ++;; since that is the only type of relocation we can use. ++(define_insn "pic_load_addr" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (unspec:SI [(match_operand:SI 1 "" "")] UNSPEC_PIC_SYM))] ++ "flag_pic && CONSTANT_POOL_ADDRESS_P(XEXP(operands[1], 0))" ++ "lddpc\t%0, %1" ++ [(set_attr "type" "load") ++ (set_attr "length" "4")] ++) ++ ++(define_insn "pic_compute_got_from_pc" ++ [(set (match_operand:SI 0 "register_operand" "+r") ++ (unspec:SI [(minus:SI (pc) ++ (match_dup 0))] UNSPEC_PIC_BASE)) ++ (use (label_ref (match_operand 1 "" "")))] ++ "flag_pic" ++ { ++ (*targetm.asm_out.internal_label) (asm_out_file, "L", ++ CODE_LABEL_NUMBER (operands[1])); ++ return \"rsub\t%0, pc\"; ++ } ++ [(set_attr "cc" "clobber") ++ (set_attr "length" "2")] ++) ++ ++;;== long long int - 64 bits ================================================== ++ ++(define_expand "movdi" ++ [(set (match_operand:DI 0 "nonimmediate_operand" "") ++ (match_operand:DI 1 "general_operand" ""))] ++ "" ++ { ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) != REG) ++ operands[1] = force_reg (DImode, operands[1]); ++ ++ }) ++ ++ ++(define_insn_and_split "*movdi_internal" ++ [(set (match_operand:DI 0 "nonimmediate_operand" "=r,r, r, r,r,r,m") ++ (match_operand:DI 1 "general_operand" "r, Ks08,Ks21,G,n,m,r"))] ++ "" ++ { ++ switch (which_alternative ){ ++ case 0: ++ case 1: ++ case 2: ++ case 3: ++ case 4: ++ return "#"; ++ case 5: ++ if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1]))) ++ return "ld.d\t%0, pc[%1 - .]"; ++ else ++ return "ld.d\t%0, %1"; ++ case 6: ++ return "st.d\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++;; Lets split all reg->reg or imm->reg transfers into two SImode transfers ++ "reload_completed && ++ (REG_P (operands[0]) && ++ (REG_P (operands[1]) ++ || GET_CODE (operands[1]) == CONST_INT ++ || GET_CODE (operands[1]) == CONST_DOUBLE))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 2) (match_dup 3))] ++ { ++ operands[2] = gen_highpart (SImode, operands[0]); ++ operands[0] = gen_lowpart (SImode, operands[0]); ++ if ( REG_P(operands[1]) ){ ++ operands[3] = gen_highpart(SImode, operands[1]); ++ operands[1] = gen_lowpart(SImode, operands[1]); ++ } else if ( GET_CODE(operands[1]) == CONST_DOUBLE ++ || GET_CODE(operands[1]) == CONST_INT ){ ++ rtx split_const[2]; ++ avr32_split_const_expr (DImode, SImode, operands[1], split_const); ++ operands[3] = split_const[1]; ++ operands[1] = split_const[0]; ++ } else { ++ internal_error("Illegal operand[1] for movdi split!"); ++ } ++ } ++ ++ [(set_attr "length" "*,*,*,*,*,4,4") ++ (set_attr "type" "*,*,*,*,*,load2,store2") ++ (set_attr "cc" "*,*,*,*,*,none,none")]) ++ ++ ++;;== 128 bits ================================================== ++(define_expand "movti" ++ [(set (match_operand:TI 0 "nonimmediate_operand" "") ++ (match_operand:TI 1 "nonimmediate_operand" ""))] ++ "" ++ { ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) != REG) ++ operands[1] = force_reg (TImode, operands[1]); ++ ++ /* We must fix any pre_dec for loads and post_inc stores */ ++ if ( GET_CODE (operands[0]) == MEM ++ && GET_CODE (XEXP(operands[0],0)) == POST_INC ){ ++ emit_move_insn(gen_rtx_MEM(TImode, XEXP(XEXP(operands[0],0),0)), operands[1]); ++ emit_insn(gen_addsi3(XEXP(XEXP(operands[0],0),0), XEXP(XEXP(operands[0],0),0), GEN_INT(GET_MODE_SIZE(TImode)))); ++ DONE; ++ } ++ ++ if ( GET_CODE (operands[1]) == MEM ++ && GET_CODE (XEXP(operands[1],0)) == PRE_DEC ){ ++ emit_insn(gen_addsi3(XEXP(XEXP(operands[1],0),0), XEXP(XEXP(operands[1],0),0), GEN_INT(-GET_MODE_SIZE(TImode)))); ++ emit_move_insn(operands[0], gen_rtx_MEM(TImode, XEXP(XEXP(operands[1],0),0))); ++ DONE; ++ } ++ }) ++ ++ ++(define_insn_and_split "*movti_internal" ++ [(set (match_operand:TI 0 "avr32_movti_dst_operand" "=r,r,,r,n,m"))] ++ "" ++ { ++ switch (which_alternative ){ ++ case 0: ++ case 3: ++ return "#"; ++ case 1: ++ return "ldm\t%p1, %0"; ++ case 2: ++ return "stm\t%p0, %1"; ++ case 4: ++ return "ld.d\t%U0, pc[%1 - .]\;ld.d\t%B0, pc[%1 - . + 8]"; ++ } ++ } ++ ++ "reload_completed && ++ (REG_P (operands[0]) && ++ (REG_P (operands[1]) ++ || (GET_CODE (operands[1]) == MEM ++ && GET_CODE (XEXP (operands[1], 0)) == SYMBOL_REF ++ && CONSTANT_POOL_ADDRESS_P (XEXP (operands[1], 0))) ++ || GET_CODE (operands[1]) == CONST_INT ++ || GET_CODE (operands[1]) == CONST_DOUBLE))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 2) (match_dup 3))] ++ { ++ operands[2] = simplify_gen_subreg ( DImode, operands[0], ++ TImode, 0 ); ++ operands[0] = simplify_gen_subreg ( DImode, operands[0], ++ TImode, 8 ); ++ if ( REG_P(operands[1]) ){ ++ operands[3] = simplify_gen_subreg ( DImode, operands[1], ++ TImode, 0 ); ++ operands[1] = simplify_gen_subreg ( DImode, operands[1], ++ TImode, 8 ); ++ } else if ( GET_CODE(operands[1]) == CONST_DOUBLE ++ || GET_CODE(operands[1]) == CONST_INT ){ ++ rtx split_const[2]; ++ avr32_split_const_expr (TImode, DImode, operands[1], split_const); ++ operands[3] = split_const[1]; ++ operands[1] = split_const[0]; ++ } else if (avr32_const_pool_ref_operand (operands[1], GET_MODE(operands[1]))){ ++ rtx split_const[2]; ++ rtx cop = avoid_constant_pool_reference (operands[1]); ++ if (operands[1] == cop) ++ cop = get_pool_constant (XEXP (operands[1], 0)); ++ avr32_split_const_expr (TImode, DImode, cop, split_const); ++ operands[3] = force_const_mem (DImode, split_const[1]); ++ operands[1] = force_const_mem (DImode, split_const[0]); ++ } else { ++ internal_error("Illegal operand[1] for movdi split!"); ++ } ++ } ++ [(set_attr "length" "*,4,4,*,8") ++ (set_attr "type" "*,load4,store4,*,load4")]) ++ ++ ++;;== float - 32 bits ========================================================== ++(define_expand "movsf" ++ [(set (match_operand:SF 0 "nonimmediate_operand" "") ++ (match_operand:SF 1 "general_operand" ""))] ++ "" ++ { ++ ++ ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) != REG) ++ operands[1] = force_reg (SFmode, operands[1]); ++ ++ }) ++ ++(define_insn "*movsf_internal" ++ [(set (match_operand:SF 0 "nonimmediate_operand" "=r,r,r,r,m") ++ (match_operand:SF 1 "general_operand" "r, G,F,m,r"))] ++ "TARGET_SOFT_FLOAT" ++ { ++ switch (which_alternative) { ++ case 0: ++ case 1: return "mov\t%0, %1"; ++ case 2: return "mov\t%0, lo(%1)\;orh\t%0, hi(%1)"; ++ case 3: ++ if ( (REG_P(XEXP(operands[1], 0)) ++ && REGNO(XEXP(operands[1], 0)) == SP_REGNUM) ++ || (GET_CODE(XEXP(operands[1], 0)) == PLUS ++ && REGNO(XEXP(XEXP(operands[1], 0), 0)) == SP_REGNUM ++ && GET_CODE(XEXP(XEXP(operands[1], 0), 1)) == CONST_INT ++ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) % 4 == 0 ++ && INTVAL(XEXP(XEXP(operands[1], 0), 1)) <= 0x1FC) ) ++ return "lddsp\t%0, %1"; ++ else if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1])) ) ++ return "lddpc\t%0, %1"; ++ else ++ return "ld.w\t%0, %1"; ++ case 4: ++ if ( (REG_P(XEXP(operands[0], 0)) ++ && REGNO(XEXP(operands[0], 0)) == SP_REGNUM) ++ || (GET_CODE(XEXP(operands[0], 0)) == PLUS ++ && REGNO(XEXP(XEXP(operands[0], 0), 0)) == SP_REGNUM ++ && GET_CODE(XEXP(XEXP(operands[0], 0), 1)) == CONST_INT ++ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) % 4 == 0 ++ && INTVAL(XEXP(XEXP(operands[0], 0), 1)) <= 0x1FC) ) ++ return "stdsp\t%0, %1"; ++ else ++ return "st.w\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ ++ [(set_attr "length" "2,4,8,4,4") ++ (set_attr "type" "alu,alu,alu2,load,store") ++ (set_attr "cc" "none,none,clobber,none,none")]) ++ ++ ++ ++;;== double - 64 bits ========================================================= ++(define_expand "movdf" ++ [(set (match_operand:DF 0 "nonimmediate_operand" "") ++ (match_operand:DF 1 "general_operand" ""))] ++ "" ++ { ++ /* One of the ops has to be in a register. */ ++ if (GET_CODE (operands[0]) != REG){ ++ operands[1] = force_reg (DFmode, operands[1]); ++ } ++ }) ++ ++ ++(define_insn_and_split "*movdf_internal" ++ [(set (match_operand:DF 0 "nonimmediate_operand" "=r,r,r,r,m") ++ (match_operand:DF 1 "general_operand" " r,G,F,m,r"))] ++ "TARGET_SOFT_FLOAT" ++ { ++ switch (which_alternative ){ ++ case 0: ++ case 1: ++ case 2: ++ return "#"; ++ case 3: ++ if ( avr32_const_pool_ref_operand(operands[1], GET_MODE(operands[1]))) ++ return "ld.d\t%0, pc[%1 - .]"; ++ else ++ return "ld.d\t%0, %1"; ++ case 4: ++ return "st.d\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ "TARGET_SOFT_FLOAT ++ && reload_completed ++ && (REG_P (operands[0]) ++ && (REG_P (operands[1]) ++ || GET_CODE (operands[1]) == CONST_DOUBLE))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 2) (match_dup 3))] ++ " ++ { ++ operands[2] = gen_highpart (SImode, operands[0]); ++ operands[0] = gen_lowpart (SImode, operands[0]); ++ operands[3] = gen_highpart(SImode, operands[1]); ++ operands[1] = gen_lowpart(SImode, operands[1]); ++ } ++ " ++ ++ [(set_attr "length" "*,*,*,4,4") ++ (set_attr "type" "*,*,*,load2,store2") ++ (set_attr "cc" "*,*,*,none,none")]) ++ ++ ++ ++ ++;;============================================================================= ++;; Move chunks of memory ++;;============================================================================= ++ ++(define_expand "movmemsi" ++ [(match_operand:BLK 0 "general_operand" "") ++ (match_operand:BLK 1 "general_operand" "") ++ (match_operand:SI 2 "const_int_operand" "") ++ (match_operand:SI 3 "const_int_operand" "")] ++ "" ++ " ++ if (avr32_gen_movmemsi (operands)) ++ DONE; ++ FAIL; ++ " ++ ++ ) ++ ++ ++ ++;;============================================================================= ++;; Bit field instructions ++;;----------------------------------------------------------------------------- ++;; Instructions to insert or extract bit-fields ++;;============================================================================= ++ ++(define_insn "insv" ++ [ (set (zero_extract:SI (match_operand:SI 0 "register_operand" "+r") ++ (match_operand:SI 1 "immediate_operand" "Ku05") ++ (match_operand:SI 2 "immediate_operand" "Ku05")) ++ (match_operand 3 "register_operand" "r"))] ++ "" ++ "bfins\t%0, %3, %2, %1" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "set_ncz")]) ++ ++ ++ ++(define_expand "extv" ++ [ (set (match_operand:SI 0 "register_operand" "") ++ (sign_extract:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "") ++ (match_operand:SI 3 "immediate_operand" "")))] ++ "" ++ { ++ if ( INTVAL(operands[2]) >= 32 ) ++ FAIL; ++ } ++) ++ ++(define_expand "extzv" ++ [ (set (match_operand:SI 0 "register_operand" "") ++ (zero_extract:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "") ++ (match_operand:SI 3 "immediate_operand" "")))] ++ "" ++ { ++ if ( INTVAL(operands[2]) >= 32 ) ++ FAIL; ++ } ++) ++ ++(define_insn "extv_internal" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (sign_extract:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku05") ++ (match_operand:SI 3 "immediate_operand" "Ku05")))] ++ "INTVAL(operands[2]) < 32" ++ "bfexts\t%0, %1, %3, %2" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "set_ncz")]) ++ ++ ++(define_insn "extzv_internal" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (zero_extract:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku05") ++ (match_operand:SI 3 "immediate_operand" "Ku05")))] ++ "INTVAL(operands[2]) < 32" ++ "bfextu\t%0, %1, %3, %2" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "set_ncz")]) ++ ++ ++ ++;;============================================================================= ++;; Some peepholes for avoiding unnecessary cast instructions ++;; followed by bfins. ++;;----------------------------------------------------------------------------- ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (zero_extend:SI (match_operand:QI 1 "register_operand" ""))) ++ (set (zero_extract:SI (match_operand 2 "register_operand" "") ++ (match_operand:SI 3 "immediate_operand" "") ++ (match_operand:SI 4 "immediate_operand" "")) ++ (match_dup 0))] ++ "((peep2_reg_dead_p(2, operands[0]) && ++ (INTVAL(operands[3]) <= 8)))" ++ [(set (zero_extract:SI (match_dup 2) ++ (match_dup 3) ++ (match_dup 4)) ++ (match_dup 1))] ++ ) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (zero_extend:SI (match_operand:HI 1 "register_operand" ""))) ++ (set (zero_extract:SI (match_operand 2 "register_operand" "") ++ (match_operand:SI 3 "immediate_operand" "") ++ (match_operand:SI 4 "immediate_operand" "")) ++ (match_dup 0))] ++ "((peep2_reg_dead_p(2, operands[0]) && ++ (INTVAL(operands[3]) <= 16)))" ++ [(set (zero_extract:SI (match_dup 2) ++ (match_dup 3) ++ (match_dup 4)) ++ (match_dup 1))] ++ ) ++ ++;;============================================================================= ++;; push bytes ++;;----------------------------------------------------------------------------- ++;; Implements the push instruction ++;;============================================================================= ++(define_insn "pushm" ++ [(set (mem:BLK (pre_dec:BLK (reg:SI SP_REGNUM))) ++ (unspec:BLK [(match_operand 0 "const_int_operand" "")] ++ UNSPEC_PUSHM))] ++ "" ++ { ++ if (INTVAL(operands[0])) { ++ return "pushm\t%r0"; ++ } else { ++ return ""; ++ } ++ } ++ [(set_attr "type" "store") ++ (set_attr "length" "2") ++ (set_attr "cc" "none")]) ++ ++(define_insn "stm" ++ [(unspec [(match_operand 0 "register_operand" "r") ++ (match_operand 1 "const_int_operand" "") ++ (match_operand 2 "const_int_operand" "")] ++ UNSPEC_STM)] ++ "" ++ { ++ if (INTVAL(operands[1])) { ++ if (INTVAL(operands[2]) != 0) ++ return "stm\t--%0, %s1"; ++ else ++ return "stm\t%0, %s1"; ++ } else { ++ return ""; ++ } ++ } ++ [(set_attr "type" "store") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++ ++ ++(define_insn "popm" ++ [(unspec [(match_operand 0 "const_int_operand" "")] ++ UNSPEC_POPM)] ++ "" ++ { ++ if (INTVAL(operands[0])) { ++ return "popm %r0"; ++ } else { ++ return ""; ++ } ++ } ++ [(set_attr "type" "load") ++ (set_attr "length" "2")]) ++ ++ ++ ++;;============================================================================= ++;; add ++;;----------------------------------------------------------------------------- ++;; Adds reg1 with reg2 and puts the result in reg0. ++;;============================================================================= ++(define_insn "add3" ++ [(set (match_operand:INTM 0 "register_operand" "=r,r,r,r,r") ++ (plus:INTM (match_operand:INTM 1 "register_operand" "%0,r,0,r,0") ++ (match_operand:INTM 2 "avr32_add_operand" "r,r,Is08,Is16,Is21")))] ++ "" ++ "@ ++ add %0, %2 ++ add %0, %1, %2 ++ sub %0, %n2 ++ sub %0, %1, %n2 ++ sub %0, %n2" ++ ++ [(set_attr "length" "2,4,2,4,4") ++ (set_attr "cc" "")]) ++ ++(define_insn "add3_lsl" ++ [(set (match_operand:INTM 0 "register_operand" "=r") ++ (plus:INTM (ashift:INTM (match_operand:INTM 1 "register_operand" "r") ++ (match_operand:INTM 3 "avr32_add_shift_immediate_operand" "Ku02")) ++ (match_operand:INTM 2 "register_operand" "r")))] ++ "" ++ "add %0, %2, %1 << %3" ++ [(set_attr "length" "4") ++ (set_attr "cc" "")]) ++ ++(define_insn "add3_lsl2" ++ [(set (match_operand:INTM 0 "register_operand" "=r") ++ (plus:INTM (match_operand:INTM 1 "register_operand" "r") ++ (ashift:INTM (match_operand:INTM 2 "register_operand" "r") ++ (match_operand:INTM 3 "avr32_add_shift_immediate_operand" "Ku02"))))] ++ "" ++ "add %0, %1, %2 << %3" ++ [(set_attr "length" "4") ++ (set_attr "cc" "")]) ++ ++ ++ ++(define_insn "add3_mul" ++ [(set (match_operand:INTM 0 "register_operand" "=r") ++ (plus:INTM (mult:INTM (match_operand:INTM 1 "register_operand" "r") ++ (match_operand:INTM 3 "immediate_operand" "Ku04" )) ++ (match_operand:INTM 2 "register_operand" "r")))] ++ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || ++ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" ++ "add %0, %2, %1 << %p3" ++ [(set_attr "length" "4") ++ (set_attr "cc" "")]) ++ ++(define_insn "add3_mul2" ++ [(set (match_operand:INTM 0 "register_operand" "=r") ++ (plus:INTM (match_operand:INTM 1 "register_operand" "r") ++ (mult:INTM (match_operand:INTM 2 "register_operand" "r") ++ (match_operand:INTM 3 "immediate_operand" "Ku04" ))))] ++ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || ++ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" ++ "add %0, %1, %2 << %p3" ++ [(set_attr "length" "4") ++ (set_attr "cc" "")]) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (ashift:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (plus:SI (match_dup 0) ++ (match_operand:SI 4 "register_operand" "")))] ++ "(peep2_reg_dead_p(2, operands[0]) && ++ (INTVAL(operands[2]) < 4 && INTVAL(operands[2]) > 0))" ++ [(set (match_dup 3) ++ (plus:SI (ashift:SI (match_dup 1) ++ (match_dup 2)) ++ (match_dup 4)))] ++ ) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (ashift:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (plus:SI (match_operand:SI 4 "register_operand" "") ++ (match_dup 0)))] ++ "(peep2_reg_dead_p(2, operands[0]) && ++ (INTVAL(operands[2]) < 4 && INTVAL(operands[2]) > 0))" ++ [(set (match_dup 3) ++ (plus:SI (ashift:SI (match_dup 1) ++ (match_dup 2)) ++ (match_dup 4)))] ++ ) ++ ++(define_insn "adddi3" ++ [(set (match_operand:DI 0 "register_operand" "=r,r") ++ (plus:DI (match_operand:DI 1 "register_operand" "%r,0") ++ (match_operand:DI 2 "register_operand" "r,r")))] ++ "" ++ "@ ++ add %0, %1, %2\;adc %m0, %m1, %m2 ++ add %0, %2\;adc %m0, %m0, %m2" ++ [(set_attr "length" "8,6") ++ (set_attr "type" "alu2") ++ (set_attr "cc" "set_vncz")]) ++ ++ ++ ++;;============================================================================= ++;; subtract ++;;----------------------------------------------------------------------------- ++;; Subtract reg2 or immediate value from reg0 and puts the result in reg0. ++;;============================================================================= ++ ++(define_insn "sub3" ++ [(set (match_operand:INTM 0 "general_operand" "=r,r,r,r,r,r,r") ++ (minus:INTM (match_operand:INTM 1 "nonmemory_operand" "0,r,0,r,0,r,Ks08") ++ (match_operand:INTM 2 "nonmemory_operand" "r,r,Ks08,Ks16,Ks21,0,r")))] ++ "" ++ "@ ++ sub %0, %2 ++ sub %0, %1, %2 ++ sub %0, %2 ++ sub %0, %1, %2 ++ sub %0, %2 ++ rsub %0, %1 ++ rsub %0, %2, %1" ++ [(set_attr "length" "2,4,2,4,4,2,4") ++ (set_attr "cc" "")]) ++ ++(define_insn "*sub3_mul" ++ [(set (match_operand:INTM 0 "register_operand" "=r,r,r") ++ (minus:INTM (match_operand:INTM 1 "register_operand" "r,0,r") ++ (mult:INTM (match_operand:INTM 2 "register_operand" "r,r,0") ++ (match_operand:SI 3 "immediate_operand" "Ku04,Ku04,Ku04" ))))] ++ "(INTVAL(operands[3]) == 0) || (INTVAL(operands[3]) == 2) || ++ (INTVAL(operands[3]) == 4) || (INTVAL(operands[3]) == 8)" ++ "@ ++ sub %0, %1, %2 << %p3 ++ sub %0, %0, %2 << %p3 ++ sub %0, %1, %0 << %p3" ++ [(set_attr "length" "4,4,4") ++ (set_attr "cc" "")]) ++ ++(define_insn "*sub3_lsl" ++ [(set (match_operand:INTM 0 "register_operand" "=r") ++ (minus:INTM (match_operand:INTM 1 "register_operand" "r") ++ (ashift:INTM (match_operand:INTM 2 "register_operand" "r") ++ (match_operand:SI 3 "avr32_add_shift_immediate_operand" "Ku02"))))] ++ "" ++ "sub %0, %1, %2 << %3" ++ [(set_attr "length" "4") ++ (set_attr "cc" "")]) ++ ++ ++(define_insn "subdi3" ++ [(set (match_operand:DI 0 "register_operand" "=r,r") ++ (minus:DI (match_operand:DI 1 "register_operand" "%r,0") ++ (match_operand:DI 2 "register_operand" "r,r")))] ++ "" ++ "@ ++ sub %0, %1, %2\;sbc %m0, %m1, %m2 ++ sub %0, %2\;sbc %m0, %m0, %m2" ++ [(set_attr "length" "8,6") ++ (set_attr "type" "alu2") ++ (set_attr "cc" "set_vncz")]) ++ ++ ++ ++;;============================================================================= ++;; multiply ++;;----------------------------------------------------------------------------- ++;; Multiply op1 and op2 and put the value in op0. ++;;============================================================================= ++ ++ ++(define_insn "mulqi3" ++ [(set (match_operand:QI 0 "register_operand" "=r,r,r") ++ (mult:QI (match_operand:QI 1 "register_operand" "%0,r,r") ++ (match_operand:QI 2 "avr32_mul_operand" "r,r,Ks08")))] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "mul %0, %2"; ++ case 1: ++ return "mul %0, %1, %2"; ++ case 2: ++ return "mul %0, %1, %2"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "type" "mulww_w,mulww_w,mulwh") ++ (set_attr "length" "2,4,4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "mulsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (mult:SI (match_operand:SI 1 "register_operand" "%0,r,r") ++ (match_operand:SI 2 "avr32_mul_operand" "r,r,Ks08")))] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "mul %0, %2"; ++ case 1: ++ return "mul %0, %1, %2"; ++ case 2: ++ return "mul %0, %1, %2"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "type" "mulww_w,mulww_w,mulwh") ++ (set_attr "length" "2,4,4") ++ (set_attr "cc" "none")]) ++ ++ ++(define_insn "mulhisi3" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (mult:SI ++ (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))))] ++ "TARGET_DSP" ++ "mulhh.w %0, %1:b, %2:b" ++ [(set_attr "type" "mulhh") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_peephole2 ++ [(match_scratch:DI 6 "r") ++ (set (match_operand:SI 0 "register_operand" "") ++ (mult:SI ++ (sign_extend:SI (match_operand:HI 1 "register_operand" "")) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "")))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (ashiftrt:SI (match_dup 0) ++ (const_int 16)))] ++ "TARGET_DSP ++ && (peep2_reg_dead_p(1, operands[0]) || (REGNO(operands[0]) == REGNO(operands[3])))" ++ [(set (match_dup 4) (sign_extend:SI (match_dup 1))) ++ (set (match_dup 6) ++ (ashift:DI (mult:DI (sign_extend:DI (match_dup 4)) ++ (sign_extend:DI (match_dup 2))) ++ (const_int 16))) ++ (set (match_dup 3) (match_dup 5))] ++ ++ "{ ++ operands[4] = gen_rtx_REG(SImode, REGNO(operands[1])); ++ operands[5] = gen_highpart (SImode, operands[4]); ++ }" ++ ) ++ ++(define_insn "mulnhisi3" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (mult:SI ++ (sign_extend:SI (neg:HI (match_operand:HI 1 "register_operand" "r"))) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))))] ++ "TARGET_DSP" ++ "mulnhh.w %0, %1:b, %2:b" ++ [(set_attr "type" "mulhh") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "machisi3" ++ [(set (match_operand:SI 0 "register_operand" "+r") ++ (plus:SI (mult:SI ++ (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) ++ (match_dup 0)))] ++ "TARGET_DSP" ++ "machh.w %0, %1:b, %2:b" ++ [(set_attr "type" "machh_w") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++ ++ ++(define_insn "mulsidi3" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (mult:DI ++ (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:SI 2 "register_operand" "r"))))] ++ "" ++ "muls.d %0, %1, %2" ++ [(set_attr "type" "mulww_d") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "umulsidi3" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (mult:DI ++ (zero_extend:DI (match_operand:SI 1 "register_operand" "%r")) ++ (zero_extend:DI (match_operand:SI 2 "register_operand" "r"))))] ++ "" ++ "mulu.d %0, %1, %2" ++ [(set_attr "type" "mulww_d") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "*mulaccsi3" ++ [(set (match_operand:SI 0 "register_operand" "+r") ++ (plus:SI (mult:SI (match_operand:SI 1 "register_operand" "%r") ++ (match_operand:SI 2 "register_operand" "r")) ++ (match_dup 0)))] ++ "" ++ "mac %0, %1, %2" ++ [(set_attr "type" "macww_w") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "mulaccsidi3" ++ [(set (match_operand:DI 0 "register_operand" "+r") ++ (plus:DI (mult:DI ++ (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:SI 2 "register_operand" "r"))) ++ (match_dup 0)))] ++ "" ++ "macs.d %0, %1, %2" ++ [(set_attr "type" "macww_d") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++(define_insn "umulaccsidi3" ++ [(set (match_operand:DI 0 "register_operand" "+r") ++ (plus:DI (mult:DI ++ (zero_extend:DI (match_operand:SI 1 "register_operand" "%r")) ++ (zero_extend:DI (match_operand:SI 2 "register_operand" "r"))) ++ (match_dup 0)))] ++ "" ++ "macu.d %0, %1, %2" ++ [(set_attr "type" "macww_d") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++ ++ ++;; Try to avoid Write-After-Write hazards for mul operations ++;; if it can be done ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (mult:SI ++ (sign_extend:SI (match_operand 1 "general_operand" "")) ++ (sign_extend:SI (match_operand 2 "general_operand" "")))) ++ (set (match_dup 0) ++ (match_operator:SI 3 "alu_operator" [(match_dup 0) ++ (match_operand 4 "general_operand" "")]))] ++ "peep2_reg_dead_p(1, operands[2])" ++ [(set (match_dup 5) ++ (mult:SI ++ (sign_extend:SI (match_dup 1)) ++ (sign_extend:SI (match_dup 2)))) ++ (set (match_dup 0) ++ (match_op_dup 3 [(match_dup 5) ++ (match_dup 4)]))] ++ "{operands[5] = gen_rtx_REG(SImode, REGNO(operands[2]));}" ++ ) ++ ++ ++ ++;;============================================================================= ++;; DSP instructions ++;;============================================================================= ++(define_insn "mulsathh_h" ++ [(set (match_operand:HI 0 "register_operand" "=r") ++ (ss_truncate:HI (ashiftrt:SI (mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 15))))] ++ "TARGET_DSP" ++ "mulsathh.h\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulhh")]) ++ ++(define_insn "mulsatrndhh_h" ++ [(set (match_operand:HI 0 "register_operand" "=r") ++ (ss_truncate:HI (ashiftrt:SI ++ (plus:SI (mult:SI (sign_extend:SI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:SI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 1073741824)) ++ (const_int 15))))] ++ "TARGET_DSP" ++ "mulsatrndhh.h\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulhh")]) ++ ++(define_insn "mulsathh_w" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ss_truncate:SI (ashift:DI (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 1))))] ++ "TARGET_DSP" ++ "mulsathh.w\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulhh")]) ++ ++(define_insn "mulsatwh_w" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ss_truncate:SI (ashiftrt:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 15))))] ++ "TARGET_DSP" ++ "mulsatwh.w\t%0, %1, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++(define_insn "mulsatrndwh_w" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ss_truncate:SI (ashiftrt:DI (plus:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 1073741824)) ++ (const_int 15))))] ++ "TARGET_DSP" ++ "mulsatrndwh.w\t%0, %1, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++(define_insn "macsathh_w" ++ [(set (match_operand:SI 0 "register_operand" "+r") ++ (plus:SI (match_dup 0) ++ (ss_truncate:SI (ashift:DI (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 1)))))] ++ "TARGET_DSP" ++ "macsathh.w\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulhh")]) ++ ++ ++(define_insn "mulwh_d" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (ashift:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 16)))] ++ "TARGET_DSP" ++ "mulwh.d\t%0, %1, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++ ++(define_insn "mulnwh_d" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (ashift:DI (mult:DI (not:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "r"))) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 16)))] ++ "TARGET_DSP" ++ "mulnwh.d\t%0, %1, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++(define_insn "macwh_d" ++ [(set (match_operand:DI 0 "register_operand" "+r") ++ (plus:DI (match_dup 0) ++ (ashift:DI (mult:DI (sign_extend:DI (match_operand:SI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r"))) ++ (const_int 16))))] ++ "TARGET_DSP" ++ "macwh.d\t%0, %1, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++(define_insn "machh_d" ++ [(set (match_operand:DI 0 "register_operand" "+r") ++ (plus:DI (match_dup 0) ++ (mult:DI (sign_extend:DI (match_operand:HI 1 "register_operand" "%r")) ++ (sign_extend:DI (match_operand:HI 2 "register_operand" "r")))))] ++ "TARGET_DSP" ++ "machh.d\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "mulwh")]) ++ ++(define_insn "satadd_w" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ss_plus:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")))] ++ "TARGET_DSP" ++ "satadd.w\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "alu_sat")]) ++ ++(define_insn "satsub_w" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ss_minus:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")))] ++ "TARGET_DSP" ++ "satsub.w\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "alu_sat")]) ++ ++(define_insn "satadd_h" ++ [(set (match_operand:HI 0 "register_operand" "=r") ++ (ss_plus:HI (match_operand:HI 1 "register_operand" "r") ++ (match_operand:HI 2 "register_operand" "r")))] ++ "TARGET_DSP" ++ "satadd.h\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "alu_sat")]) ++ ++(define_insn "satsub_h" ++ [(set (match_operand:HI 0 "register_operand" "=r") ++ (ss_minus:HI (match_operand:HI 1 "register_operand" "r") ++ (match_operand:HI 2 "register_operand" "r")))] ++ "TARGET_DSP" ++ "satsub.h\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none") ++ (set_attr "type" "alu_sat")]) ++ ++ ++;;============================================================================= ++;; smin ++;;----------------------------------------------------------------------------- ++;; Set reg0 to the smallest value of reg1 and reg2. It is used for signed ++;; values in the registers. ++;;============================================================================= ++(define_insn "sminsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (smin:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")))] ++ "" ++ "min %0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++;;============================================================================= ++;; smax ++;;----------------------------------------------------------------------------- ++;; Set reg0 to the largest value of reg1 and reg2. It is used for signed ++;; values in the registers. ++;;============================================================================= ++(define_insn "smaxsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (smax:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")))] ++ "" ++ "max %0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++ ++;;============================================================================= ++;; Logical operations ++;;----------------------------------------------------------------------------- ++ ++;; Split up simple DImode logical operations. Simply perform the logical ++;; operation on the upper and lower halves of the registers. ++(define_split ++ [(set (match_operand:DI 0 "register_operand" "") ++ (match_operator:DI 6 "logical_binary_operator" ++ [(match_operand:DI 1 "register_operand" "") ++ (match_operand:DI 2 "register_operand" "")]))] ++ "reload_completed" ++ [(set (match_dup 0) (match_op_dup:SI 6 [(match_dup 1) (match_dup 2)])) ++ (set (match_dup 3) (match_op_dup:SI 6 [(match_dup 4) (match_dup 5)]))] ++ " ++ { ++ operands[3] = gen_highpart (SImode, operands[0]); ++ operands[0] = gen_lowpart (SImode, operands[0]); ++ operands[4] = gen_highpart (SImode, operands[1]); ++ operands[1] = gen_lowpart (SImode, operands[1]); ++ operands[5] = gen_highpart (SImode, operands[2]); ++ operands[2] = gen_lowpart (SImode, operands[2]); ++ }" ++) ++ ++;;============================================================================= ++;; Logical operations with shifted operand ++;;============================================================================= ++(define_insn "si_lshift" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (logical:SI (match_operator:SI 4 "logical_shift_operator" ++ [(match_operand:SI 2 "register_operand" "r") ++ (match_operand:SI 3 "immediate_operand" "Ku05")]) ++ (match_operand:SI 1 "register_operand" "r")))] ++ "" ++ { ++ if ( GET_CODE(operands[4]) == ASHIFT ) ++ return "\t%0, %1, %2 << %3"; ++ else ++ return "\t%0, %1, %2 >> %3"; ++ } ++ ++ [(set_attr "cc" "set_z")] ++) ++ ++ ++;;************************************************ ++;; Peepholes for detecting logical operantions ++;; with shifted operands ++;;************************************************ ++ ++(define_peephole ++ [(set (match_operand:SI 3 "register_operand" "") ++ (match_operator:SI 5 "logical_shift_operator" ++ [(match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "")])) ++ (set (match_operand:SI 0 "register_operand" "") ++ (logical:SI (match_operand:SI 4 "register_operand" "") ++ (match_dup 3)))] ++ "(dead_or_set_p(insn, operands[3])) || (REGNO(operands[3]) == REGNO(operands[0]))" ++ { ++ if ( GET_CODE(operands[5]) == ASHIFT ) ++ return "\t%0, %4, %1 << %2"; ++ else ++ return "\t%0, %4, %1 >> %2"; ++ } ++ [(set_attr "cc" "set_z")] ++ ) ++ ++(define_peephole ++ [(set (match_operand:SI 3 "register_operand" "") ++ (match_operator:SI 5 "logical_shift_operator" ++ [(match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "")])) ++ (set (match_operand:SI 0 "register_operand" "") ++ (logical:SI (match_dup 3) ++ (match_operand:SI 4 "register_operand" "")))] ++ "(dead_or_set_p(insn, operands[3])) || (REGNO(operands[3]) == REGNO(operands[0]))" ++ { ++ if ( GET_CODE(operands[5]) == ASHIFT ) ++ return "\t%0, %4, %1 << %2"; ++ else ++ return "\t%0, %4, %1 >> %2"; ++ } ++ [(set_attr "cc" "set_z")] ++ ) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (match_operator:SI 5 "logical_shift_operator" ++ [(match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "")])) ++ (set (match_operand:SI 3 "register_operand" "") ++ (logical:SI (match_operand:SI 4 "register_operand" "") ++ (match_dup 0)))] ++ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[3]) == REGNO(operands[0]))" ++ ++ [(set (match_dup 3) ++ (logical:SI (match_op_dup:SI 5 [(match_dup 1) (match_dup 2)]) ++ (match_dup 4)))] ++ ++ "" ++) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (match_operator:SI 5 "logical_shift_operator" ++ [(match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" "")])) ++ (set (match_operand:SI 3 "register_operand" "") ++ (logical:SI (match_dup 0) ++ (match_operand:SI 4 "register_operand" "")))] ++ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[3]) == REGNO(operands[0]))" ++ ++ [(set (match_dup 3) ++ (logical:SI (match_op_dup:SI 5 [(match_dup 1) (match_dup 2)]) ++ (match_dup 4)))] ++ ++ "" ++) ++ ++ ++;;============================================================================= ++;; and ++;;----------------------------------------------------------------------------- ++;; Store the result after a bitwise logical-and between reg0 and reg2 in reg0. ++;;============================================================================= ++ ++(define_insn "andnsi" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (and:SI (match_operand:SI 1 "register_operand" "0") ++ (not:SI (match_operand:SI 2 "register_operand" "r"))))] ++ "" ++ "andn %0, %2" ++ [(set_attr "cc" "set_z") ++ (set_attr "length" "2")] ++) ++ ++ ++ ++ ++(define_insn "andsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r, r, r, r") ++ (and:SI (match_operand:SI 1 "register_operand" "%0, r, 0, r") ++ (match_operand:SI 2 "nonmemory_operand" "r, M, i, r")))] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "and\t%0, %2"; ++ case 1: ++ { ++ int i, first_set = -1; ++ /* Search for first bit set in mask */ ++ for ( i = 31; i >= 0; --i ) ++ if ( INTVAL(operands[2]) & (1 << i) ){ ++ first_set = i; ++ break; ++ } ++ operands[2] = gen_rtx_CONST_INT(SImode, first_set + 1); ++ return "bfextu\t%0, %1, 0, %2"; ++ } ++ case 2: ++ if ( one_bit_cleared_operand(operands[2], VOIDmode) ){ ++ int bitpos; ++ for ( bitpos = 0; bitpos < 32; bitpos++ ) ++ if ( !(INTVAL(operands[2]) & (1 << bitpos)) ) ++ break; ++ operands[2] = gen_rtx_CONST_INT(SImode, bitpos); ++ return "cbr\t%0, %2"; ++ } else if ( (INTVAL(operands[2]) >= 0) && ++ (INTVAL(operands[2]) <= 65535) ) ++ return "andl\t%0, %2, COH"; ++ else if ( (INTVAL(operands[2]) < 0) && ++ (INTVAL(operands[2]) >= -65536 ) ) ++ return "andl\t%0, lo(%2)"; ++ else if ( ((INTVAL(operands[2]) & 0xffff) == 0xffff) ) ++ return "andh\t%0, hi(%2)"; ++ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) ++ return "andh\t%0, hi(%2), COH"; ++ else ++ return "andh\t%0, hi(%2)\;andl\t%0, lo(%2)"; ++ case 3: ++ return "and\t%0, %1, %2"; ++ default: ++ abort(); ++ } ++ } ++ ++ [(set_attr "length" "2,4,8,4") ++ (set_attr "cc" "set_z")]) ++ ++ ++(define_insn "anddi3" ++ [(set (match_operand:DI 0 "register_operand" "=&r,&r") ++ (and:DI (match_operand:DI 1 "register_operand" "%0,r") ++ (match_operand:DI 2 "register_operand" "r,r")))] ++ "" ++ "#" ++ [(set_attr "length" "8") ++ (set_attr "cc" "clobber")] ++) ++ ++;;============================================================================= ++;; or ++;;----------------------------------------------------------------------------- ++;; Store the result after a bitwise inclusive-or between reg0 and reg2 in reg0. ++;;============================================================================= ++ ++(define_insn "iorsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (ior:SI (match_operand:SI 1 "register_operand" "%0,0,r" ) ++ (match_operand:SI 2 "nonmemory_operand" "r ,i,r")))] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "or\t%0, %2"; ++ case 1: ++ if ( one_bit_set_operand(operands[2], VOIDmode) ){ ++ int bitpos; ++ for (bitpos = 0; bitpos < 32; bitpos++) ++ if (INTVAL(operands[2]) & (1 << bitpos)) ++ break; ++ operands[2] = gen_rtx_CONST_INT( SImode, bitpos); ++ return "sbr\t%0, %2"; ++ } else if ( (INTVAL(operands[2]) >= 0) && ++ (INTVAL(operands[2]) <= 65535) ) ++ return "orl\t%0, %2"; ++ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) ++ return "orh\t%0, hi(%2)"; ++ else ++ return "orh\t%0, hi(%2)\;orl\t%0, lo(%2)"; ++ case 2: ++ return "or\t%0, %1, %2"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,8,4") ++ (set_attr "cc" "set_z")]) ++ ++ ++;(define_insn "iorsi3" ++; [(set (match_operand:SI 0 "register_operand" "=r, r, r") ++; (ior:SI (match_operand:SI 1 "avr32_logical_insn_operand" "r, r, rA" ) ++; (match_operand:SI 2 "register_operand" "0, i, r")))] ++; "" ++; { ++; switch (which_alternative){ ++; case 0: ++; return "or %0, %2"; ++; case 1: ++; if ( one_bit_set_operand(operands[2], VOIDmode) ){ ++; int i, bitpos; ++; for ( i = 0; i < 32; i++ ) ++; if ( INTVAL(operands[2]) & (1 << i) ){ ++; bitpos = i; ++; break; ++; } ++; operands[2] = gen_rtx_CONST_INT( SImode, bitpos); ++; return "sbr %0, %2"; ++; } else if ( (INTVAL(operands[2]) >= 0) && ++; (INTVAL(operands[2]) <= 65535) ) ++; return "orl %0, %2"; ++; else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) ++; return "orh %0, hi(%2)"; ++; else ++; return "orh %0, hi(%2)\;orl %0, lo(%2)"; ++; case 2: ++; return "or %0, %2, %1"; ++; } ++; } ++; [(set_attr "length" "2,8,4") ++; (set_attr "cc" "set_z")]) ++ ++(define_insn "iordi3" ++ [(set (match_operand:DI 0 "register_operand" "=&r,&r") ++ (ior:DI (match_operand:DI 1 "register_operand" "%0,r") ++ (match_operand:DI 2 "register_operand" "r,r")))] ++ "" ++ "#" ++ [(set_attr "length" "8") ++ (set_attr "cc" "clobber")] ++) ++ ++;;============================================================================= ++;; xor bytes ++;;----------------------------------------------------------------------------- ++;; Store the result after a bitwise exclusive-or between reg0 and reg2 in reg0. ++;;============================================================================= ++ ++(define_insn "xorsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (xor:SI (match_operand:SI 1 "register_operand" "0,0,r") ++ (match_operand:SI 2 "nonmemory_operand" "r,i,r")))] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "eor %0, %2"; ++ case 1: ++ if ( (INTVAL(operands[2]) >= 0) && ++ (INTVAL(operands[2]) <= 65535) ) ++ return "eorl %0, %2"; ++ else if ( ((INTVAL(operands[2]) & 0xffff) == 0x0) ) ++ return "eorh %0, hi(%2)"; ++ else ++ return "eorh %0, hi(%2)\;eorl %0, lo(%2)"; ++ case 2: ++ return "eor %0, %1, %2"; ++ default: ++ abort(); ++ } ++ } ++ ++ [(set_attr "length" "2,8,4") ++ (set_attr "cc" "set_z")]) ++ ++(define_insn "xordi3" ++ [(set (match_operand:DI 0 "register_operand" "=&r,&r") ++ (xor:DI (match_operand:DI 1 "register_operand" "%0,r") ++ (match_operand:DI 2 "register_operand" "r,r")))] ++ "" ++ "#" ++ [(set_attr "length" "8") ++ (set_attr "cc" "clobber")] ++) ++ ++;;============================================================================= ++;; divmod ++;;----------------------------------------------------------------------------- ++;; Signed division that produces both a quotient and a remainder. ++;;============================================================================= ++(define_expand "divmodsi4" ++ [(parallel [ ++ (parallel [ ++ (set (match_operand:SI 0 "register_operand" "=r") ++ (div:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r"))) ++ (set (match_operand:SI 3 "register_operand" "=r") ++ (mod:SI (match_dup 1) ++ (match_dup 2)))]) ++ (use (match_dup 4))])] ++ "" ++ { ++ if (! no_new_pseudos) { ++ operands[4] = gen_reg_rtx (DImode); ++ ++ emit_insn(gen_divmodsi4_internal(operands[4],operands[1],operands[2])); ++ emit_move_insn(operands[0], gen_rtx_SUBREG( SImode, operands[4], 4)); ++ emit_move_insn(operands[3], gen_rtx_SUBREG( SImode, operands[4], 0)); ++ ++ DONE; ++ } else { ++ FAIL; ++ } ++ ++ }) ++ ++ ++(define_insn "divmodsi4_internal" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (unspec:DI [(match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")] ++ UNSPEC_DIVMODSI4_INTERNAL))] ++ "" ++ "divs %0, %1, %2" ++ [(set_attr "type" "div") ++ (set_attr "cc" "none")]) ++ ++ ++;;============================================================================= ++;; udivmod ++;;----------------------------------------------------------------------------- ++;; Unsigned division that produces both a quotient and a remainder. ++;;============================================================================= ++(define_expand "udivmodsi4" ++ [(parallel [ ++ (parallel [ ++ (set (match_operand:SI 0 "register_operand" "=r") ++ (udiv:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r"))) ++ (set (match_operand:SI 3 "register_operand" "=r") ++ (umod:SI (match_dup 1) ++ (match_dup 2)))]) ++ (use (match_dup 4))])] ++ "" ++ { ++ if (! no_new_pseudos) { ++ operands[4] = gen_reg_rtx (DImode); ++ ++ emit_insn(gen_udivmodsi4_internal(operands[4],operands[1],operands[2])); ++ emit_move_insn(operands[0], gen_rtx_SUBREG( SImode, operands[4], 4)); ++ emit_move_insn(operands[3], gen_rtx_SUBREG( SImode, operands[4], 0)); ++ ++ DONE; ++ } else { ++ FAIL; ++ } ++ }) ++ ++(define_insn "udivmodsi4_internal" ++ [(set (match_operand:DI 0 "register_operand" "=r") ++ (unspec:DI [(match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "register_operand" "r")] ++ UNSPEC_UDIVMODSI4_INTERNAL))] ++ "" ++ "divu %0, %1, %2" ++ [(set_attr "type" "div") ++ (set_attr "cc" "none")]) ++ ++ ++;;============================================================================= ++;; Arithmetic-shift left ++;;----------------------------------------------------------------------------- ++;; Arithmetic-shift reg0 left by reg2 or immediate value. ++;;============================================================================= ++ ++(define_insn "ashlsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (ashift:SI (match_operand:SI 1 "register_operand" "r,0,r") ++ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] ++ "" ++ "@ ++ lsl %0, %1, %2 ++ lsl %0, %2 ++ lsl %0, %1, %2" ++ [(set_attr "length" "4,2,4") ++ (set_attr "cc" "set_ncz")]) ++ ++;;============================================================================= ++;; Arithmetic-shift right ++;;----------------------------------------------------------------------------- ++;; Arithmetic-shift reg0 right by an immediate value. ++;;============================================================================= ++ ++(define_insn "ashrsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (ashiftrt:SI (match_operand:SI 1 "register_operand" "r,0,r") ++ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] ++ "" ++ "@ ++ asr %0, %1, %2 ++ asr %0, %2 ++ asr %0, %1, %2" ++ [(set_attr "length" "4,2,4") ++ (set_attr "cc" "set_ncz")]) ++ ++;;============================================================================= ++;; Logical shift right ++;;----------------------------------------------------------------------------- ++;; Logical shift reg0 right by an immediate value. ++;;============================================================================= ++ ++(define_insn "lshrsi3" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r") ++ (lshiftrt:SI (match_operand:SI 1 "register_operand" "r,0,r") ++ (match_operand:SI 2 "nonmemory_operand" "r,Ku05,Ku05")))] ++ "" ++ "@ ++ lsr %0, %1, %2 ++ lsr %0, %2 ++ lsr %0, %1, %2" ++ [(set_attr "length" "4,2,4") ++ (set_attr "cc" "set_ncz")]) ++ ++ ++;;============================================================================= ++;; neg ++;;----------------------------------------------------------------------------- ++;; Negate operand 1 and store the result in operand 0. ++;;============================================================================= ++(define_insn "negsi2" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (neg:SI (match_operand:SI 1 "register_operand" "0")))] ++ "" ++ "neg %0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "set_vncz")]) ++ ++;;============================================================================= ++;; abs ++;;----------------------------------------------------------------------------- ++;; Store the absolute value of operand 1 into operand 0. ++;;============================================================================= ++(define_insn "abssi2" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (abs:SI (match_operand:SI 1 "register_operand" "0")))] ++ "" ++ "abs %0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "set_z")]) ++ ++ ++;;============================================================================= ++;; one_cmpl ++;;----------------------------------------------------------------------------- ++;; Store the bitwise-complement of operand 1 into operand 0. ++;;============================================================================= ++ ++(define_insn "one_cmplsi2" ++ [(set (match_operand:SI 0 "register_operand" "=r,r") ++ (not:SI (match_operand:SI 1 "register_operand" "r,0")))] ++ "" ++ "@ ++ rsub %0, %1, -1 ++ com %0" ++ [(set_attr "length" "4,2") ++ (set_attr "cc" "set_z")]) ++ ++ ++;;============================================================================= ++;; Bit load ++;;----------------------------------------------------------------------------- ++;; Load a bit into Z and C flags ++;;============================================================================= ++(define_insn "bldsi" ++ [(set (cc0) ++ (and:SI (match_operand:SI 0 "register_operand" "r") ++ (match_operand:SI 1 "one_bit_set_operand" "i")))] ++ "" ++ "bld\t%0, %p1" ++ [(set_attr "length" "4") ++ (set_attr "cc" "bld")] ++ ) ++ ++ ++;;============================================================================= ++;; Compare ++;;----------------------------------------------------------------------------- ++;; Compare reg0 with reg1 or an immediate value. ++;;============================================================================= ++ ++(define_expand "cmpqi" ++ [(set (cc0) ++ (compare:QI ++ (match_operand:QI 0 "general_operand" "") ++ (match_operand:QI 1 "general_operand" "")))] ++ "" ++ "{ ++ ++ if ( GET_CODE(operands[0]) != REG ++ && GET_CODE(operands[0]) != SUBREG) ++ operands[0] = force_reg(QImode, operands[0]); ++ ++ ++ if ( GET_CODE(operands[1]) != REG ++ && GET_CODE(operands[1]) != SUBREG ) ++ operands[1] = force_reg(QImode, operands[1]); ++ ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ emit_insn(gen_cmpqi_internal(operands[0], operands[1])); ++ DONE; ++ }" ++) ++ ++ ++(define_insn "cmpqi_internal" ++ [(set (cc0) ++ (compare:QI ++ (match_operand:QI 0 "register_operand" "r") ++ (match_operand:QI 1 "register_operand" "r")))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), QImode, operands[0], operands[1])); ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "cc" "compare")]) ++ ++(define_expand "cmphi" ++ [(set (cc0) ++ (compare:HI ++ (match_operand:HI 0 "register_operand" "") ++ (match_operand:HI 1 "register_operand" "")))] ++ "" ++ "{ ++ ++ //if ( (GET_CODE(operands[0]) == REG ++ // || GET_CODE(operands[0]) == SUBREG) ++ // && (GET_CODE(operands[1]) == CONST_INT ++ // && avr32_const_ok_for_constraint_p (INTVAL(operands[1]), 'K', \"Ks21\")) ){ ++ // operands[0] = convert_to_mode(SImode, operands[0], 0); ++ // avr32_compare_op0 = operands[0]; ++ // avr32_compare_op1 = operands[1]; ++ // emit_insn(gen_cmpsi_internal(operands[0], operands[1])); ++ // DONE; ++ //} ++ ++ if ( GET_CODE(operands[0]) != REG ++ && GET_CODE(operands[0]) != SUBREG ) ++ operands[0] = force_reg(HImode, operands[0]); ++ ++ ++ if ( GET_CODE(operands[1]) != REG ++ && GET_CODE(operands[1]) != SUBREG) ++ operands[1] = force_reg(HImode, operands[1]); ++ ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ }" ++) ++ ++ ++(define_insn "cmphi_internal" ++ [(set (cc0) ++ (compare:HI ++ (match_operand:HI 0 "register_operand" "r") ++ (match_operand:HI 1 "register_operand" "r")))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), HImode, operands[0], operands[1])); ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "cc" "compare")]) ++ ++ ++(define_expand "cmpsi" ++ [(set (cc0) ++ (compare:SI ++ (match_operand:SI 0 "register_operand" "") ++ (match_operand:SI 1 "register_immediate_operand" "")))] ++ "" ++ "{ ++ if ( GET_CODE(operands[0]) != REG ++ && GET_CODE(operands[0]) != SUBREG ) ++ operands[0] = force_reg(SImode, operands[0]); ++ ++ if ( GET_CODE(operands[1]) != REG ++ && GET_CODE(operands[1]) != SUBREG ++ && GET_CODE(operands[1]) != CONST_INT ) ++ operands[1] = force_reg(SImode, operands[1]); ++ ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ ++ }" ++) ++ ++ ++(define_insn "cmpsi_internal" ++ [(set (cc0) ++ (compare:SI ++ (match_operand:SI 0 "register_operand" "r, r, r") ++ (match_operand:SI 1 "register_immediate_operand" "r, Ks06, Ks21")))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), SImode, operands[0], operands[1])); ++ return ""; ++ } ++ ++ [(set_attr "length" "2,2,4") ++ (set_attr "cc" "compare")]) ++ ++ ++(define_expand "cmpdi" ++ [(set (cc0) ++ (compare:DI ++ (match_operand:DI 0 "register_operand" "") ++ (match_operand:DI 1 "register_immediate_operand" "")))] ++ "" ++ { ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ } ++) ++ ++(define_insn "cmpdi_internal" ++ [(set (cc0) ++ (compare:DI ++ (match_operand:DI 0 "register_operand" "r") ++ (match_operand:DI 1 "register_immediate_operand" "rKu20")))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), DImode, operands[0], operands[1])); ++ return ""; ++ } ++ ++ [(set_attr "length" "6") ++ (set_attr "type" "alu2") ++ (set_attr "cc" "compare")]) ++ ++ ++ ++;;============================================================================= ++;; Test if zero ++;;----------------------------------------------------------------------------- ++;; Compare reg against zero and set the condition codes. ++;;============================================================================= ++ ++ ++(define_expand "tstsi" ++ [(set (cc0) ++ (match_operand:SI 0 "register_operand" ""))] ++ "" ++ { ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = const0_rtx; ++ } ++) ++ ++(define_insn "tstsi_internal" ++ [(set (cc0) ++ (match_operand:SI 0 "register_operand" "r"))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), SImode, operands[0], const0_rtx)); ++ ++ return ""; ++ } ++ [(set_attr "length" "2") ++ (set_attr "cc" "compare")]) ++ ++ ++(define_expand "tstdi" ++ [(set (cc0) ++ (match_operand:DI 0 "register_operand" ""))] ++ "" ++ { ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = const0_rtx; ++ } ++) ++ ++(define_insn "tstdi_internal" ++ [(set (cc0) ++ (match_operand:DI 0 "register_operand" "r"))] ++ "" ++ { ++ set_next_insn_cond(insn, ++ avr32_output_cmp(get_next_insn_cond(insn), DImode, operands[0], const0_rtx)); ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "type" "alu2") ++ (set_attr "cc" "compare")]) ++ ++ ++ ++;;============================================================================= ++;; Convert operands ++;;----------------------------------------------------------------------------- ++;; ++;;============================================================================= ++(define_insn "truncdisi2" ++ [(set (match_operand:SI 0 "general_operand" "") ++ (truncate:SI (match_operand:DI 1 "general_operand" "")))] ++ "" ++ "truncdisi2") ++ ++;;============================================================================= ++;; Extend ++;;----------------------------------------------------------------------------- ++;; ++;;============================================================================= ++ ++ ++(define_insn "extendhisi2" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") ++ (sign_extend:SI (match_operand:HI 1 "nonimmediate_operand" "0,r,,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "casts.h\t%0"; ++ case 1: ++ return "bfexts\t%0, %1, 0, 16"; ++ case 2: ++ case 3: ++ return "ld.sh\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz,set_ncz,none,none") ++ (set_attr "type" "alu,alu,load_rm,load_rm")]) ++ ++(define_insn "extendqisi2" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") ++ (sign_extend:SI (match_operand:QI 1 "extendqi_operand" "0,r,RKu00,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "casts.b\t%0"; ++ case 1: ++ return "bfexts\t%0, %1, 0, 8"; ++ case 2: ++ case 3: ++ return "ld.sb\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz,set_ncz,none,none") ++ (set_attr "type" "alu,alu,load_rm,load_rm")]) ++ ++(define_insn "extendqihi2" ++ [(set (match_operand:HI 0 "register_operand" "=r,r,r,r") ++ (sign_extend:HI (match_operand:QI 1 "extendqi_operand" "0,r,RKu00,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "casts.b\t%0"; ++ case 1: ++ return "bfexts\t%0, %1, 0, 8"; ++ case 2: ++ case 3: ++ return "ld.sb\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz,set_ncz,none,none") ++ (set_attr "type" "alu,alu,load_rm,load_rm")]) ++ ++ ++;;============================================================================= ++;; Zero-extend ++;;----------------------------------------------------------------------------- ++;; ++;;============================================================================= ++ ++(define_insn "zero_extendhisi2" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") ++ (zero_extend:SI (match_operand:HI 1 "nonimmediate_operand" "0,r,,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "castu.h\t%0"; ++ case 1: ++ return "bfextu\t%0, %1, 0, 16"; ++ case 2: ++ case 3: ++ return "ld.uh\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz,set_ncz,none,none") ++ (set_attr "type" "alu,alu,load_rm,load_rm")]) ++ ++(define_insn "zero_extendqisi2" ++ [(set (match_operand:SI 0 "register_operand" "=r,r,r,r") ++ (zero_extend:SI (match_operand:QI 1 "nonimmediate_operand" "0,r,,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "castu.b\t%0"; ++ case 1: ++ return "bfextu\t%0, %1, 0, 8"; ++ case 2: ++ case 3: ++ return "ld.ub\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz, set_ncz, none, none") ++ (set_attr "type" "alu, alu, load_rm, load_rm")]) ++ ++(define_insn "zero_extendqihi2" ++ [(set (match_operand:HI 0 "register_operand" "=r,r,r,r") ++ (zero_extend:HI (match_operand:QI 1 "nonimmediate_operand" "0,r,,m")))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ return "castu.b\t%0"; ++ case 1: ++ return "bfextu\t%0, %1, 0, 8"; ++ case 2: ++ case 3: ++ return "ld.ub\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,2,4") ++ (set_attr "cc" "set_ncz, set_ncz, none, none") ++ (set_attr "type" "alu, alu, load_rm, load_rm")]) ++ ++ ++ ++;;============================================================================= ++;; Conditional set register ++;; sr{cond4} rd ++;;----------------------------------------------------------------------------- ++ ++;;Because of the same issue as with conditional moves and adds we must ++;;not separate the compare instrcution from the scc instruction as ++;;they might be sheduled "badly". ++ ++(define_insn "s" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (any_cond:SI (cc0) ++ (const_int 0)))] ++ "" ++ "sr\t%0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "none")]) ++ ++(define_insn "smi" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (unspec:SI [(cc0) ++ (const_int 0)] UNSPEC_COND_MI))] ++ "" ++ "srmi\t%0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "none")]) ++ ++(define_insn "spl" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (unspec:SI [(cc0) ++ (const_int 0)] UNSPEC_COND_PL))] ++ "" ++ "srpl\t%0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "none")]) ++ ++ ++;;============================================================================= ++;; Conditional branch ++;;----------------------------------------------------------------------------- ++;; Branch to label if the specified condition codes are set. ++;;============================================================================= ++; branch if negative ++(define_insn "bmi" ++ [(set (pc) ++ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_MI) ++ (label_ref (match_operand 0 "" "")) ++ (pc)))] ++ "" ++ "brmi %0" ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++(define_insn "*bmi-reverse" ++ [(set (pc) ++ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_MI) ++ (pc) ++ (label_ref (match_operand 0 "" ""))))] ++ "" ++ "brpl %0" ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++; branch if positive ++(define_insn "bpl" ++ [(set (pc) ++ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_PL) ++ (label_ref (match_operand 0 "" "")) ++ (pc)))] ++ "" ++ "brpl %0" ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++(define_insn "*bpl-reverse" ++ [(set (pc) ++ (if_then_else (unspec:CC [(cc0) (const_int 0)] UNSPEC_COND_PL) ++ (pc) ++ (label_ref (match_operand 0 "" ""))))] ++ "" ++ "brmi %0" ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++; branch if equal ++(define_insn "b" ++ [(set (pc) ++ (if_then_else (any_cond:CC (cc0) ++ (const_int 0)) ++ (label_ref (match_operand 0 "" "")) ++ (pc)))] ++ "" ++ "br %0 " ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++ ++(define_insn "*b-reverse" ++ [(set (pc) ++ (if_then_else (any_cond:CC (cc0) ++ (const_int 0)) ++ (pc) ++ (label_ref (match_operand 0 "" ""))))] ++ "" ++ "br %0 " ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 254)) ++ (le (minus (pc) (match_dup 0)) (const_int 256))) ++ (const_int 2)] ; use compact branch ++ (const_int 4))) ; use extended branch ++ (set_attr "cc" "none")]) ++ ++ ++ ++;============================================================================= ++; Conditional Add/Subtract ++;----------------------------------------------------------------------------- ++; sub{cond4} Rd, imm ++;============================================================================= ++ ++ ++(define_expand "addcc" ++ [(set (match_operand:ADDCC 0 "register_operand" "") ++ (if_then_else:ADDCC (match_operator 1 "avr32_comparison_operator" ++ [(match_dup 4) ++ (match_dup 5)]) ++ (match_operand:ADDCC 2 "register_operand" "") ++ (plus:ADDCC ++ (match_dup 2) ++ (match_operand:ADDCC 3 "avr32_cond_immediate_operand" ""))))] ++ "" ++ { ++ /* Delete compare instruction as it is merged into this instruction */ ++ remove_insn (get_last_insn_anywhere ()); ++ ++ operands[4] = avr32_compare_op0; ++ operands[5] = avr32_compare_op1; ++ } ++ ) ++ ++ ++(define_insn "addcc_cmp" ++ [(set (match_operand:ADDCC 0 "register_operand" "=r") ++ (if_then_else:ADDCC (match_operator 1 "avr32_comparison_operator" ++ [(match_operand:CMP 4 "register_operand" "r") ++ (match_operand:CMP 5 "" "")]) ++ (match_operand:ADDCC 2 "register_operand" "0") ++ (plus:ADDCC ++ (match_dup 2) ++ (match_operand:ADDCC 3 "avr32_cond_immediate_operand" "Is08"))))] ++ "" ++ { ++ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); ++ ++ return "sub%i1\t%0, -%3"; ++ } ++ [(set_attr "length" "8") ++ (set_attr "cc" "cmp_cond_insn")]) ++ ++;============================================================================= ++; Conditional Move ++;----------------------------------------------------------------------------- ++; mov{cond4} Rd, (Rs/imm) ++;============================================================================= ++(define_expand "movcc" ++ [(set (match_operand:MOVCC 0 "register_operand" "") ++ (if_then_else:MOVCC (match_operator 1 "avr32_comparison_operator" ++ [(match_dup 4) ++ (match_dup 5)]) ++ (match_operand:MOVCC 2 "avr32_cond_register_immediate_operand" "") ++ (match_operand:MOVCC 3 "avr32_cond_register_immediate_operand" "")))] ++ "" ++ { ++ /* Delete compare instruction as it is merged into this instruction */ ++ remove_insn (get_last_insn_anywhere ()); ++ ++ operands[4] = avr32_compare_op0; ++ operands[5] = avr32_compare_op1; ++ } ++ ) ++ ++ ++(define_insn "movcc_cmp" ++ [(set (match_operand:MOVCC 0 "register_operand" "=r,r,r") ++ (if_then_else:MOVCC (match_operator 1 "avr32_comparison_operator" ++ [(match_operand:CMP 4 "register_operand" "r,r,r") ++ (match_operand:CMP 5 "" ",,")]) ++ (match_operand:MOVCC 2 "avr32_cond_register_immediate_operand" "0, rKs08,rKs08") ++ (match_operand:MOVCC 3 "avr32_cond_register_immediate_operand" "rKs08,0,rKs08")))] ++ "" ++ { ++ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); ++ ++ switch( which_alternative ){ ++ case 0: ++ return "mov%i1 %0, %3"; ++ case 1: ++ return "mov%1 %0, %2"; ++ case 2: ++ return "mov%1 %0, %2\;mov%i1 %0, %3"; ++ default: ++ abort(); ++ } ++ ++ ++ } ++ [(set_attr "length" "8,8,12") ++ (set_attr "cc" "cmp_cond_insn")]) ++ ++ ++;;============================================================================= ++;; jump ++;;----------------------------------------------------------------------------- ++;; Jump inside a function; an unconditional branch to a label. ++;;============================================================================= ++(define_insn "jump" ++ [(set (pc) ++ (label_ref (match_operand 0 "" "")))] ++ "" ++ { ++ if (get_attr_length(insn) > 4) ++ return "Can't jump this far"; ++ return (get_attr_length(insn) == 2 ? ++ "rjmp %0" : "bral %0"); ++ } ++ [(set_attr "type" "branch") ++ (set (attr "length") ++ (cond [(and (le (minus (match_dup 0) (pc)) (const_int 1022)) ++ (le (minus (pc) (match_dup 0)) (const_int 1024))) ++ (const_int 2) ; use rjmp ++ (le (match_dup 0) (const_int 1048575)) ++ (const_int 4)] ; use bral ++ (const_int 8))) ; do something else ++ (set_attr "cc" "none")]) ++ ++;;============================================================================= ++;; call ++;;----------------------------------------------------------------------------- ++;; Subroutine call instruction returning no value. ++;;============================================================================= ++(define_insn "call_internal" ++ [(parallel [(call (mem:SI (match_operand:SI 0 "avr32_call_operand" "r,U,T,W")) ++ (match_operand 1 "" "")) ++ (clobber (reg:SI LR_REGNUM))])] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "icall\t%0"; ++ case 1: ++ return "rcall\t%0"; ++ case 2: ++ return "mcall\t%0"; ++ case 3: ++ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) ++ return "call\t%0"; ++ else ++ return "mcall\tr6[%0@got]"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "type" "call") ++ (set_attr "length" "2,4,4,10") ++ (set_attr "cc" "clobber")]) ++ ++ ++(define_expand "call" ++ [(parallel [(call (match_operand:SI 0 "" "") ++ (match_operand 1 "" "")) ++ (clobber (reg:SI LR_REGNUM))])] ++ "" ++ { ++ rtx call_address; ++ if ( GET_CODE(operands[0]) != MEM ) ++ FAIL; ++ ++ call_address = XEXP(operands[0], 0); ++ ++ /* If assembler supports call pseudo insn and the call ++ address is a symbol then nothing special needs to be done. */ ++ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ++ && (GET_CODE(call_address) == SYMBOL_REF) ){ ++ /* We must however mark the function as using the GOT if ++ flag_pic is set, since the call insn might turn into ++ a mcall using the GOT ptr register. */ ++ if ( flag_pic ){ ++ current_function_uses_pic_offset_table = 1; ++ emit_call_insn(gen_call_internal(call_address, operands[1])); ++ DONE; ++ } ++ } else { ++ if ( flag_pic && ++ GET_CODE(call_address) == SYMBOL_REF ){ ++ current_function_uses_pic_offset_table = 1; ++ emit_call_insn(gen_call_internal(call_address, operands[1])); ++ DONE; ++ } ++ ++ if ( !SYMBOL_REF_RCALL_FUNCTION_P(operands[0]) ){ ++ if ( optimize_size && ++ GET_CODE(call_address) == SYMBOL_REF ){ ++ call_address = force_const_mem(SImode, call_address); ++ } else { ++ call_address = force_reg(SImode, call_address); ++ } ++ } ++ } ++ emit_call_insn(gen_call_internal(call_address, operands[1])); ++ DONE; ++ } ++) ++ ++;;============================================================================= ++;; call_value ++;;----------------------------------------------------------------------------- ++;; Subrutine call instruction returning a value. ++;;============================================================================= ++(define_expand "call_value" ++ [(parallel [(set (match_operand:SI 0 "" "") ++ (call (match_operand:SI 1 "" "") ++ (match_operand 2 "" ""))) ++ (clobber (reg:SI LR_REGNUM))])] ++ "" ++ { ++ rtx call_address; ++ if ( GET_CODE(operands[1]) != MEM ) ++ FAIL; ++ ++ call_address = XEXP(operands[1], 0); ++ ++ /* If assembler supports call pseudo insn and the call ++ address is a symbol then nothing special needs to be done. */ ++ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ++ && (GET_CODE(call_address) == SYMBOL_REF) ){ ++ /* We must however mark the function as using the GOT if ++ flag_pic is set, since the call insn might turn into ++ a mcall using the GOT ptr register. */ ++ if ( flag_pic ) { ++ current_function_uses_pic_offset_table = 1; ++ emit_call_insn(gen_call_value_internal(operands[0], call_address, operands[2])); ++ DONE; ++ } ++ } else { ++ if ( flag_pic && ++ GET_CODE(call_address) == SYMBOL_REF ){ ++ current_function_uses_pic_offset_table = 1; ++ emit_call_insn(gen_call_value_internal(operands[0], call_address, operands[2])); ++ DONE; ++ } ++ ++ if ( !SYMBOL_REF_RCALL_FUNCTION_P(operands[1]) ){ ++ if ( optimize_size && ++ GET_CODE(call_address) == SYMBOL_REF){ ++ call_address = force_const_mem(SImode, call_address); ++ } else { ++ call_address = force_reg(SImode, call_address); ++ } ++ } ++ } ++ emit_call_insn(gen_call_value_internal(operands[0], call_address, ++ operands[2])); ++ DONE; ++ ++ }) ++ ++(define_insn "call_value_internal" ++ [(parallel [(set (match_operand 0 "register_operand" "=r,r,r,r") ++ (call (mem:SI (match_operand:SI 1 "avr32_call_operand" "r,U,T,W")) ++ (match_operand 2 "" ""))) ++ (clobber (reg:SI LR_REGNUM))])] ++ ;; Operand 2 not used on the AVR32. ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "icall\t%1"; ++ case 1: ++ return "rcall\t%1"; ++ case 2: ++ return "mcall\t%1"; ++ case 3: ++ if ( TARGET_HAS_ASM_ADDR_PSEUDOS ) ++ return "call\t%1"; ++ else ++ return "mcall\tr6[%1@got]"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "type" "call") ++ (set_attr "length" "2,4,4,10") ++ (set_attr "cc" "call_set")]) ++ ++ ++;;============================================================================= ++;; untyped_call ++;;----------------------------------------------------------------------------- ++;; Subrutine call instruction returning a value of any type. ++;; The code is copied from m68k.md (except gen_blockage is removed) ++;; Fixme! ++;;============================================================================= ++(define_expand "untyped_call" ++ [(parallel [(call (match_operand 0 "avr32_call_operand" "") ++ (const_int 0)) ++ (match_operand 1 "" "") ++ (match_operand 2 "" "")])] ++ "" ++ { ++ int i; ++ ++ emit_call_insn (GEN_CALL (operands[0], const0_rtx, NULL, const0_rtx)); ++ ++ for (i = 0; i < XVECLEN (operands[2], 0); i++) { ++ rtx set = XVECEXP (operands[2], 0, i); ++ emit_move_insn (SET_DEST (set), SET_SRC (set)); ++ } ++ ++ /* The optimizer does not know that the call sets the function value ++ registers we stored in the result block. We avoid problems by ++ claiming that all hard registers are used and clobbered at this ++ point. */ ++ emit_insn (gen_blockage ()); ++ ++ DONE; ++ }) ++ ++ ++;;============================================================================= ++;; return ++;;============================================================================= ++ ++(define_insn "return" ++ [(return)] ++ "USE_RETURN_INSN (FALSE)" ++ { ++ avr32_output_return_instruction(TRUE, FALSE, NULL, NULL); ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "type" "call")] ++ ) ++ ++(define_insn "return_cond" ++ [(set (pc) ++ (if_then_else (match_operand 0 "avr32_comparison_operand" "") ++ (return) ++ (pc)))] ++ "USE_RETURN_INSN (TRUE)" ++ "ret%0\tr12"; ++ [(set_attr "type" "call")]) ++ ++ ++(define_insn "return_imm" ++ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) ++ (use (reg RETVAL_REGNUM)) ++ (return)])] ++ "USE_RETURN_INSN (FALSE) && ++ ((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" ++ { ++ avr32_output_return_instruction(TRUE, FALSE, NULL, operands[0]); ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "type" "call")] ++ ) ++ ++(define_insn "return_imm_cond" ++ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) ++ (use (reg RETVAL_REGNUM)) ++ (set (pc) ++ (if_then_else (match_operand 1 "avr32_comparison_operand" "") ++ (return) ++ (pc)))])] ++ "USE_RETURN_INSN (TRUE) && ++ ((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" ++ "ret%1\t%0"; ++ [(set_attr "type" "call")] ++ ) ++ ++(define_insn "return_reg" ++ [(set (reg RETVAL_REGNUM) (match_operand:MOVM 0 "register_operand" "r")) ++ (use (reg RETVAL_REGNUM)) ++ (return)] ++ "USE_RETURN_INSN (TRUE)" ++ "retal %0"; ++ [(set_attr "type" "call")] ++ ) ++ ++(define_insn "return_reg_cond" ++ [(set (reg RETVAL_REGNUM) (match_operand:MOVM 0 "register_operand" "r")) ++ (use (reg RETVAL_REGNUM)) ++ (set (pc) ++ (if_then_else (match_operator 1 "avr32_comparison_operator" ++ [(cc0) (const_int 0)]) ++ (return) ++ (pc)))] ++ "USE_RETURN_INSN (TRUE)" ++ "ret%1\t%0"; ++ [(set_attr "type" "call")]) ++ ++;;============================================================================= ++;; nop ++;;----------------------------------------------------------------------------- ++;; No-op instruction. ++;;============================================================================= ++(define_insn "nop" ++ [(const_int 0)] ++ "" ++ "nop" ++ [(set_attr "length" "2") ++ (set_attr "type" "alu") ++ (set_attr "cc" "none")]) ++ ++;;============================================================================= ++;; nonlocal_goto_receiver ++;;----------------------------------------------------------------------------- ++;; For targets with a return stack we must make sure to flush the return stack ++;; since it will be corrupt after a nonlocal goto. ++;;============================================================================= ++(define_expand "nonlocal_goto_receiver" ++ [(const_int 0)] ++ "TARGET_RETURN_STACK" ++ " ++ { ++ emit_insn ( gen_frs() ); ++ DONE; ++ } ++ " ++ ) ++ ++ ++;;============================================================================= ++;; builtin_setjmp_receiver ++;;----------------------------------------------------------------------------- ++;; For pic code we need to reload the pic register. ++;; For targets with a return stack we must make sure to flush the return stack ++;; since it will probably be corrupted. ++;;============================================================================= ++(define_expand "builtin_setjmp_receiver" ++ [(label_ref (match_operand 0 "" ""))] ++ "flag_pic" ++ " ++ { ++ if ( TARGET_RETURN_STACK ) ++ emit_insn ( gen_frs() ); ++ avr32_load_pic_register (); ++ DONE; ++ } ++ " ++) ++ ++ ++;;============================================================================= ++;; indirect_jump ++;;----------------------------------------------------------------------------- ++;; Jump to an address in reg or memory. ++;;============================================================================= ++(define_expand "indirect_jump" ++ [(set (pc) ++ (match_operand:SI 0 "general_operand" ""))] ++ "" ++ { ++ /* One of the ops has to be in a register. */ ++ if ( (flag_pic || TARGET_HAS_ASM_ADDR_PSEUDOS ) ++ && !avr32_legitimate_pic_operand_p(operands[0]) ) ++ operands[0] = legitimize_pic_address (operands[0], SImode, 0); ++ else if ( flag_pic && avr32_address_operand(operands[0], GET_MODE(operands[0])) ) ++ /* If we have an address operand then this function uses the pic register. */ ++ current_function_uses_pic_offset_table = 1; ++ }) ++ ++ ++(define_insn "indirect_jump_internal" ++ [(set (pc) ++ (match_operand:SI 0 "general_operand" "r,m,W"))] ++ "" ++ { ++ switch( which_alternative ){ ++ case 0: ++ return "mov\tpc, %0"; ++ case 1: ++ if ( avr32_const_pool_ref_operand(operands[0], GET_MODE(operands[0])) ) ++ return "lddpc\tpc, %0"; ++ else ++ return "ld.w\tpc, %0"; ++ case 2: ++ if ( flag_pic ) ++ return "ld.w\tpc, r6[%0@got]"; ++ else ++ return "lda.w\tpc, %0"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "2,4,8") ++ (set_attr "type" "call,call,call") ++ (set_attr "cc" "none,none,clobber")]) ++ ++ ++;;============================================================================= ++;; casesi and tablejump ++;;============================================================================= ++(define_insn "tablejump_add" ++ [(set (pc) ++ (plus:SI (match_operand:SI 0 "register_operand" "r") ++ (mult:SI (match_operand:SI 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku04" )))) ++ (use (label_ref (match_operand 3 "" "")))] ++ "flag_pic && ++ ((INTVAL(operands[2]) == 0) || (INTVAL(operands[2]) == 2) || ++ (INTVAL(operands[2]) == 4) || (INTVAL(operands[2]) == 8))" ++ "add\tpc, %0, %1 << %p2" ++ [(set_attr "length" "4") ++ (set_attr "cc" "clobber")]) ++ ++(define_insn "tablejump_insn" ++ [(set (pc) (match_operand:SI 0 "memory_operand" "m")) ++ (use (label_ref (match_operand 1 "" "")))] ++ "!flag_pic" ++ "ld.w\tpc, %0" ++ [(set_attr "length" "4") ++ (set_attr "type" "call") ++ (set_attr "cc" "none")]) ++ ++(define_expand "casesi" ++ [(match_operand:SI 0 "register_operand" "") ; index to jump on ++ (match_operand:SI 1 "const_int_operand" "") ; lower bound ++ (match_operand:SI 2 "const_int_operand" "") ; total range ++ (match_operand:SI 3 "" "") ; table label ++ (match_operand:SI 4 "" "")] ; Out of range label ++ "" ++ " ++ { ++ rtx reg; ++ rtx index = operands[0]; ++ rtx low_bound = operands[1]; ++ rtx range = operands[2]; ++ rtx table_label = operands[3]; ++ rtx oor_label = operands[4]; ++ ++ if (low_bound != const0_rtx) ++ { ++ if (!avr32_const_ok_for_constraint_p(INTVAL (low_bound), 'I', \"Is21\")){ ++ reg = force_reg(SImode, GEN_INT (INTVAL (low_bound))); ++ emit_insn (gen_subsi3 (reg, index, ++ reg)); ++ } else { ++ reg = gen_reg_rtx (SImode); ++ emit_insn (gen_addsi3 (reg, index, ++ GEN_INT (-INTVAL (low_bound)))); ++ } ++ index = reg; ++ } ++ ++ if (!avr32_const_ok_for_constraint_p (INTVAL (range), 'K', \"Ks21\")) ++ range = force_reg (SImode, range); ++ ++ emit_cmp_and_jump_insns ( index, range, GTU, NULL_RTX, SImode, 1, oor_label ); ++ reg = gen_reg_rtx (SImode); ++ emit_move_insn ( reg, gen_rtx_LABEL_REF (VOIDmode, table_label)); ++ ++ if ( flag_pic ) ++ emit_jump_insn ( gen_tablejump_add ( reg, index, GEN_INT(4), table_label)); ++ else ++ emit_jump_insn ( ++ gen_tablejump_insn ( gen_rtx_MEM ( SImode, ++ gen_rtx_PLUS ( SImode, ++ reg, ++ gen_rtx_MULT ( SImode, ++ index, ++ GEN_INT(4)))), ++ table_label)); ++ DONE; ++ }" ++) ++ ++ ++ ++(define_insn "prefetch" ++ [(prefetch (match_operand:SI 0 "register_operand" "r") ++ (match_operand 1 "const_int_operand" "") ++ (match_operand 2 "const_int_operand" ""))] ++ "" ++ { ++ return "pref\t%0[0]"; ++ } ++ ++ [(set_attr "length" "4") ++ (set_attr "type" "load") ++ (set_attr "cc" "none")]) ++ ++ ++ ++;;============================================================================= ++;; prologue ++;;----------------------------------------------------------------------------- ++;; This pattern, if defined, emits RTL for entry to a function. The function ++;; entry i responsible for setting up the stack frame, initializing the frame ++;; pointer register, saving callee saved registers, etc. ++;;============================================================================= ++(define_expand "prologue" ++ [(clobber (const_int 0))] ++ "" ++ " ++ avr32_expand_prologue(); ++ DONE; ++ " ++ ) ++ ++;;============================================================================= ++;; eh_return ++;;----------------------------------------------------------------------------- ++;; This pattern, if defined, affects the way __builtin_eh_return, and ++;; thence the call frame exception handling library routines, are ++;; built. It is intended to handle non-trivial actions needed along ++;; the abnormal return path. ++;; ++;; The address of the exception handler to which the function should ++;; return is passed as operand to this pattern. It will normally need ++;; to copied by the pattern to some special register or memory ++;; location. If the pattern needs to determine the location of the ++;; target call frame in order to do so, it may use ++;; EH_RETURN_STACKADJ_RTX, if defined; it will have already been ++;; assigned. ++;; ++;; If this pattern is not defined, the default action will be to ++;; simply copy the return address to EH_RETURN_HANDLER_RTX. Either ++;; that macro or this pattern needs to be defined if call frame ++;; exception handling is to be used. ++ ++;; We can't expand this before we know where the link register is stored. ++(define_insn_and_split "eh_return" ++ [(unspec_volatile [(match_operand:SI 0 "register_operand" "r")] ++ VUNSPEC_EH_RETURN) ++ (clobber (match_scratch:SI 1 "=&r"))] ++ "" ++ "#" ++ "reload_completed" ++ [(const_int 0)] ++ " ++ { ++ avr32_set_return_address (operands[0], operands[1]); ++ DONE; ++ }" ++ ) ++ ++;;============================================================================= ++;; ffssi2 ++;;----------------------------------------------------------------------------- ++(define_insn "ffssi2" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (ffs:SI (match_operand:SI 1 "register_operand" "r"))) ] ++ "" ++ "mov %0, %1 ++ brev %0 ++ clz %0, %0 ++ sub %0, -1 ++ cp %0, 33 ++ moveq %0, 0" ++ [(set_attr "length" "18") ++ (set_attr "cc" "clobber")] ++ ) ++ ++ ++ ++;;============================================================================= ++;; swap_h ++;;----------------------------------------------------------------------------- ++(define_insn "*swap_h" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (ior:SI (ashift:SI (match_dup 0) (const_int 16)) ++ (lshiftrt:SI (match_dup 0) (const_int 16))))] ++ "" ++ "swap.h %0" ++ [(set_attr "length" "2")] ++ ) ++ ++(define_insn_and_split "bswap_16" ++ [ (set (match_operand:HI 0 "avr32_bswap_operand" "=r,RKs13,r") ++ (ior:HI (and:HI (lshiftrt:HI (match_operand:HI 1 "avr32_bswap_operand" "r,r,RKs13") ++ (const_int 8)) ++ (const_int 255)) ++ (ashift:HI (and:HI (match_dup 1) ++ (const_int 255)) ++ (const_int 8))))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ if ( REGNO(operands[0]) == REGNO(operands[1])) ++ return "swap.bh\t%0"; ++ else ++ return "mov\t%0, %1\;swap.bh\t%0"; ++ case 1: ++ return "stswp.h\t%0, %1"; ++ case 2: ++ return "ldswp.sh\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ ++ "(reload_completed && ++ REG_P(operands[0]) && REG_P(operands[1]) ++ && (REGNO(operands[0]) != REGNO(operands[1])))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 0) ++ (ior:HI (and:HI (lshiftrt:HI (match_dup 0) ++ (const_int 8)) ++ (const_int 255)) ++ (ashift:HI (and:HI (match_dup 0) ++ (const_int 255)) ++ (const_int 8))))] ++ "" ++ ++ [(set_attr "length" "4,4,4") ++ (set_attr "type" "alu,store,load_rm")] ++ ) ++ ++(define_insn_and_split "bswap_32" ++ [ (set (match_operand:SI 0 "avr32_bswap_operand" "=r,RKs14,r") ++ (ior:SI (ior:SI (lshiftrt:SI (and:SI (match_operand:SI 1 "avr32_bswap_operand" "=r,r,RKs14") ++ (const_int 4278190080)) ++ (const_int 24)) ++ (lshiftrt:SI (and:SI (match_dup 1) ++ (const_int 16711680)) ++ (const_int 8))) ++ (ior:SI (ashift:SI (and:SI (match_dup 1) ++ (const_int 65280)) ++ (const_int 8)) ++ (ashift:SI (and:SI (match_dup 1) ++ (const_int 255)) ++ (const_int 24)))))] ++ "" ++ { ++ switch ( which_alternative ){ ++ case 0: ++ if ( REGNO(operands[0]) == REGNO(operands[1])) ++ return "swap.b\t%0"; ++ else ++ return "mov\t%0, %1\;swap.b\t%0"; ++ case 1: ++ return "stswp.w\t%0, %1"; ++ case 2: ++ return "ldswp.w\t%0, %1"; ++ default: ++ abort(); ++ } ++ } ++ "(reload_completed && ++ REG_P(operands[0]) && REG_P(operands[1]) ++ && (REGNO(operands[0]) != REGNO(operands[1])))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 0) ++ (ior:SI (ior:SI (lshiftrt:SI (and:SI (match_dup 0) ++ (const_int 4278190080)) ++ (const_int 24)) ++ (lshiftrt:SI (and:SI (match_dup 0) ++ (const_int 16711680)) ++ (const_int 8))) ++ (ior:SI (ashift:SI (and:SI (match_dup 0) ++ (const_int 65280)) ++ (const_int 8)) ++ (ashift:SI (and:SI (match_dup 0) ++ (const_int 255)) ++ (const_int 24)))))] ++ "" ++ ++ [(set_attr "length" "4,4,4") ++ (set_attr "type" "alu,store,load_rm")] ++ ) ++ ++ ++;;============================================================================= ++;; blockage ++;;----------------------------------------------------------------------------- ++;; UNSPEC_VOLATILE is considered to use and clobber all hard registers and ++;; all of memory. This blocks insns from being moved across this point. ++ ++(define_insn "blockage" ++ [(unspec_volatile [(const_int 0)] VUNSPEC_BLOCKAGE)] ++ "" ++ "" ++ [(set_attr "length" "0")] ++) ++ ++;;============================================================================= ++;; clzsi2 ++;;----------------------------------------------------------------------------- ++(define_insn "clzsi2" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (clz:SI (match_operand:SI 1 "register_operand" "r"))) ] ++ "" ++ "clz %0, %1" ++ [(set_attr "length" "4") ++ (set_attr "cc" "set_z")] ++ ) ++ ++;;============================================================================= ++;; ctzsi2 ++;;----------------------------------------------------------------------------- ++(define_insn "ctzsi2" ++ [ (set (match_operand:SI 0 "register_operand" "=r,r") ++ (ctz:SI (match_operand:SI 1 "register_operand" "0,r"))) ] ++ "" ++ "@ ++ brev\t%0\;clz\t%0, %0 ++ mov\t%0, %1\;brev\t%0\;clz\t%0, %0" ++ [(set_attr "length" "8") ++ (set_attr "cc" "set_z")] ++ ) ++ ++;;============================================================================= ++;; cache instructions ++;;----------------------------------------------------------------------------- ++(define_insn "cache" ++ [ (unspec_volatile [(match_operand:SI 0 "register_operand" "r") ++ (match_operand:SI 1 "immediate_operand" "Ku05")] VUNSPEC_CACHE)] ++ "" ++ "cache %0[0], %1" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "sync" ++ [ (unspec_volatile [(match_operand:SI 0 "immediate_operand" "Ku08")] VUNSPEC_SYNC)] ++ "" ++ "sync %0" ++ [(set_attr "length" "4")] ++ ) ++ ++;;============================================================================= ++;; TLB instructions ++;;----------------------------------------------------------------------------- ++(define_insn "tlbr" ++ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBR)] ++ "" ++ "tlbr" ++ [(set_attr "length" "2")] ++ ) ++ ++(define_insn "tlbw" ++ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBW)] ++ "" ++ "tlbw" ++ [(set_attr "length" "2")] ++ ) ++ ++(define_insn "tlbs" ++ [ (unspec_volatile [(const_int 0)] VUNSPEC_TLBS)] ++ "" ++ "tlbs" ++ [(set_attr "length" "2")] ++ ) ++ ++;;============================================================================= ++;; Breakpoint instruction ++;;----------------------------------------------------------------------------- ++(define_insn "breakpoint" ++ [ (unspec_volatile [(const_int 0)] VUNSPEC_BREAKPOINT)] ++ "" ++ "breakpoint" ++ [(set_attr "length" "2")] ++ ) ++ ++ ++;;============================================================================= ++;; mtsr/mfsr instruction ++;;----------------------------------------------------------------------------- ++(define_insn "mtsr" ++ [ (unspec_volatile [(match_operand 0 "immediate_operand" "i") ++ (match_operand:SI 1 "register_operand" "r")] VUNSPEC_MTSR)] ++ "" ++ "mtsr\t%0, %1" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mfsr" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "i")] VUNSPEC_MFSR)) ] ++ "" ++ "mfsr\t%0, %1" ++ [(set_attr "length" "4")] ++ ) ++ ++;;============================================================================= ++;; mtdr/mfdr instruction ++;;----------------------------------------------------------------------------- ++(define_insn "mtdr" ++ [ (unspec_volatile [(match_operand 0 "immediate_operand" "i") ++ (match_operand:SI 1 "register_operand" "r")] VUNSPEC_MTDR)] ++ "" ++ "mtdr\t%0, %1" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mfdr" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "i")] VUNSPEC_MFDR)) ] ++ "" ++ "mfdr\t%0, %1" ++ [(set_attr "length" "4")] ++ ) ++ ++;;============================================================================= ++;; musfr ++;;----------------------------------------------------------------------------- ++(define_insn "musfr" ++ [ (unspec_volatile [(match_operand:SI 0 "register_operand" "r")] VUNSPEC_MUSFR)] ++ "" ++ "musfr\t%0" ++ [(set_attr "length" "2") ++ (set_attr "cc" "clobber")] ++ ) ++ ++(define_insn "mustr" ++ [ (set (match_operand:SI 0 "register_operand" "=r") ++ (unspec_volatile:SI [(const_int 0)] VUNSPEC_MUSTR)) ] ++ "" ++ "mustr\t%0" ++ [(set_attr "length" "2")] ++ ) ++ ++;;============================================================================= ++;; Flush Return Stack instruction ++;;----------------------------------------------------------------------------- ++(define_insn "frs" ++ [ (unspec_volatile [(const_int 0)] VUNSPEC_FRS)] ++ "" ++ "frs" ++ [(set_attr "length" "2") ++ (set_attr "cc" "none")] ++ ) ++ ++ ++;;============================================================================= ++;; Saturation Round Scale instruction ++;;----------------------------------------------------------------------------- ++(define_insn "sats" ++ [ (set (match_operand:SI 0 "register_operand" "+r") ++ (unspec:SI [(match_dup 0) ++ (match_operand 1 "immediate_operand" "Ku05") ++ (match_operand 2 "immediate_operand" "Ku05")] ++ UNSPEC_SATS)) ] ++ "TARGET_DSP" ++ "sats\t%0 >> %1, %2" ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4")] ++ ) ++ ++(define_insn "satu" ++ [ (set (match_operand:SI 0 "register_operand" "+r") ++ (unspec:SI [(match_dup 0) ++ (match_operand 1 "immediate_operand" "Ku05") ++ (match_operand 2 "immediate_operand" "Ku05")] ++ UNSPEC_SATU)) ] ++ "TARGET_DSP" ++ "satu\t%0 >> %1, %2" ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4")] ++ ) ++ ++(define_insn "satrnds" ++ [ (set (match_operand:SI 0 "register_operand" "+r") ++ (unspec:SI [(match_dup 0) ++ (match_operand 1 "immediate_operand" "Ku05") ++ (match_operand 2 "immediate_operand" "Ku05")] ++ UNSPEC_SATRNDS)) ] ++ "TARGET_DSP" ++ "satrnds\t%0 >> %1, %2" ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4")] ++ ) ++ ++(define_insn "satrndu" ++ [ (set (match_operand:SI 0 "register_operand" "+r") ++ (unspec:SI [(match_dup 0) ++ (match_operand 1 "immediate_operand" "Ku05") ++ (match_operand 2 "immediate_operand" "Ku05")] ++ UNSPEC_SATRNDU)) ] ++ "TARGET_DSP" ++ "sats\t%0 >> %1, %2" ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4")] ++ ) ++ ++;; Special patterns for dealing with the constant pool ++ ++(define_insn "align_4" ++ [(unspec_volatile [(const_int 0)] VUNSPEC_ALIGN)] ++ "" ++ { ++ assemble_align (32); ++ return ""; ++ } ++ [(set_attr "length" "2")] ++) ++ ++(define_insn "consttable_start" ++ [(unspec_volatile [(const_int 0)] VUNSPEC_POOL_START)] ++ "" ++ { ++ return ".cpool"; ++ } ++ [(set_attr "length" "0")] ++ ) ++ ++(define_insn "consttable_end" ++ [(unspec_volatile [(const_int 0)] VUNSPEC_POOL_END)] ++ "" ++ { ++ making_const_table = FALSE; ++ return ""; ++ } ++ [(set_attr "length" "0")] ++) ++ ++ ++(define_insn "consttable_4" ++ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_4)] ++ "" ++ { ++ making_const_table = TRUE; ++ switch (GET_MODE_CLASS (GET_MODE (operands[0]))) ++ { ++ case MODE_FLOAT: ++ { ++ REAL_VALUE_TYPE r; ++ char real_string[1024]; ++ REAL_VALUE_FROM_CONST_DOUBLE (r, operands[0]); ++ real_to_decimal(real_string, &r, 1024, 0, 1); ++ asm_fprintf (asm_out_file, "\t.float\t%s\n", real_string); ++ break; ++ } ++ default: ++ assemble_integer (operands[0], 4, 0, 1); ++ break; ++ } ++ return ""; ++ } ++ [(set_attr "length" "4")] ++) ++ ++(define_insn "consttable_8" ++ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_8)] ++ "" ++ { ++ making_const_table = TRUE; ++ switch (GET_MODE_CLASS (GET_MODE (operands[0]))) ++ { ++ case MODE_FLOAT: ++ { ++ REAL_VALUE_TYPE r; ++ char real_string[1024]; ++ REAL_VALUE_FROM_CONST_DOUBLE (r, operands[0]); ++ real_to_decimal(real_string, &r, 1024, 0, 1); ++ asm_fprintf (asm_out_file, "\t.double\t%s\n", real_string); ++ break; ++ } ++ default: ++ assemble_integer(operands[0], 8, 0, 1); ++ break; ++ } ++ return ""; ++ } ++ [(set_attr "length" "8")] ++) ++ ++(define_insn "consttable_16" ++ [(unspec_volatile [(match_operand 0 "" "")] VUNSPEC_POOL_16)] ++ "" ++ { ++ making_const_table = TRUE; ++ assemble_integer(operands[0], 16, 0, 1); ++ return ""; ++ } ++ [(set_attr "length" "16")] ++) ++ ++;;============================================================================= ++;; coprocessor instructions ++;;----------------------------------------------------------------------------- ++(define_insn "cop" ++ [ (unspec_volatile [(match_operand 0 "immediate_operand" "Ku03") ++ (match_operand 1 "immediate_operand" "Ku04") ++ (match_operand 2 "immediate_operand" "Ku04") ++ (match_operand 3 "immediate_operand" "Ku04") ++ (match_operand 4 "immediate_operand" "Ku07")] VUNSPEC_COP)] ++ "" ++ "cop\tcp%0, cr%1, cr%2, cr%3, %4" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mvcrsi" ++ [ (set (match_operand:SI 0 "avr32_cop_move_operand" "=r,<,Z") ++ (unspec_volatile:SI [(match_operand 1 "immediate_operand" "Ku03,Ku03,Ku03") ++ (match_operand 2 "immediate_operand" "Ku04,Ku04,Ku04")] ++ VUNSPEC_MVCR)) ] ++ "" ++ "@ ++ mvcr.w\tcp%1, %0, cr%2 ++ stcm.w\tcp%1, %0, cr%2 ++ stc.w\tcp%1, %0, cr%2" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mvcrdi" ++ [ (set (match_operand:DI 0 "avr32_cop_move_operand" "=r,<,Z") ++ (unspec_volatile:DI [(match_operand 1 "immediate_operand" "Ku03,Ku03,Ku03") ++ (match_operand 2 "immediate_operand" "Ku04,Ku04,Ku04")] ++ VUNSPEC_MVCR)) ] ++ "" ++ "@ ++ mvcr.d\tcp%1, %0, cr%2 ++ stcm.d\tcp%1, %0, cr%2-cr%i2 ++ stc.d\tcp%1, %0, cr%2" ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mvrcsi" ++ [ (unspec_volatile:SI [(match_operand 0 "immediate_operand" "Ku03,Ku03,Ku03") ++ (match_operand 1 "immediate_operand" "Ku04,Ku04,Ku04") ++ (match_operand:SI 2 "avr32_cop_move_operand" "r,>,Z")] ++ VUNSPEC_MVRC)] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "mvrc.w\tcp%0, cr%1, %2"; ++ case 1: ++ return "ldcm.w\tcp%0, %2, cr%1"; ++ case 2: ++ return "ldc.w\tcp%0, cr%1, %2"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "4")] ++ ) ++ ++(define_insn "mvrcdi" ++ [ (unspec_volatile:DI [(match_operand 0 "immediate_operand" "Ku03,Ku03,Ku03") ++ (match_operand 1 "immediate_operand" "Ku04,Ku04,Ku04") ++ (match_operand:DI 2 "avr32_cop_move_operand" "r,>,Z")] ++ VUNSPEC_MVRC)] ++ "" ++ { ++ switch (which_alternative){ ++ case 0: ++ return "mvrc.d\tcp%0, cr%1, %2"; ++ case 1: ++ return "ldcm.d\tcp%0, %2, cr%1-cr%i1"; ++ case 2: ++ return "ldc.d\tcp%0, cr%1, %2"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "length" "4")] ++ ) ++ ++;;============================================================================= ++;; epilogue ++;;----------------------------------------------------------------------------- ++;; This pattern emits RTL for exit from a function. The function exit is ++;; responsible for deallocating the stack frame, restoring callee saved ++;; registers and emitting the return instruction. ++;; ToDo: using TARGET_ASM_FUNCTION_PROLOGUE instead. ++;;============================================================================= ++(define_expand "epilogue" ++ [(unspec_volatile [(return)] VUNSPEC_EPILOGUE)] ++ "" ++ " ++ if (USE_RETURN_INSN (FALSE)){ ++ emit_jump_insn (gen_return ()); ++ DONE; ++ } ++ emit_jump_insn (gen_rtx_UNSPEC_VOLATILE (VOIDmode, ++ gen_rtvec (1, ++ gen_rtx_RETURN (VOIDmode)), ++ VUNSPEC_EPILOGUE)); ++ DONE; ++ " ++ ) ++ ++(define_insn "*epilogue_insns" ++ [(unspec_volatile [(return)] VUNSPEC_EPILOGUE)] ++ "" ++ { ++ avr32_output_return_instruction (FALSE, FALSE, NULL, NULL); ++ return ""; ++ } ++ ; Length is absolute worst case ++ [(set_attr "type" "branch") ++ (set_attr "length" "12")] ++ ) ++ ++(define_insn "*epilogue_insns_ret_imm" ++ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) ++ (use (reg RETVAL_REGNUM)) ++ (unspec_volatile [(return)] VUNSPEC_EPILOGUE)])] ++ "((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" ++ { ++ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[0]); ++ return ""; ++ } ++ ; Length is absolute worst case ++ [(set_attr "type" "branch") ++ (set_attr "length" "12")] ++ ) ++ ++(define_insn "sibcall_epilogue" ++ [(unspec_volatile [(const_int 0)] VUNSPEC_EPILOGUE)] ++ "" ++ { ++ avr32_output_return_instruction (FALSE, FALSE, NULL, NULL); ++ return ""; ++ } ++;; Length is absolute worst case ++ [(set_attr "type" "branch") ++ (set_attr "length" "12")] ++ ) ++ ++(define_insn "*sibcall_epilogue_insns_ret_imm" ++ [(parallel [(set (reg RETVAL_REGNUM) (match_operand 0 "immediate_operand" "i")) ++ (use (reg RETVAL_REGNUM)) ++ (unspec_volatile [(const_int 0)] VUNSPEC_EPILOGUE)])] ++ "((INTVAL(operands[0]) == -1) || (INTVAL(operands[0]) == 0) || (INTVAL(operands[0]) == 1))" ++ { ++ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[0]); ++ return ""; ++ } ++ ; Length is absolute worst case ++ [(set_attr "type" "branch") ++ (set_attr "length" "12")] ++ ) ++ ++(define_insn "ldxi" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (mem:SI (plus:SI ++ (match_operand:SI 1 "register_operand" "r") ++ (mult:SI (zero_extract:SI (match_operand:SI 2 "register_operand" "r") ++ (const_int 8) ++ (match_operand:SI 3 "immediate_operand" "Ku05")) ++ (const_int 4)))))] ++ "(INTVAL(operands[3]) == 24 || INTVAL(operands[3]) == 16 || INTVAL(operands[3]) == 8 ++ || INTVAL(operands[3]) == 0)" ++ { ++ switch ( INTVAL(operands[3]) ){ ++ case 0: ++ return "ld.w %0, %1[%2:b << 2]"; ++ case 8: ++ return "ld.w %0, %1[%2:l << 2]"; ++ case 16: ++ return "ld.w %0, %1[%2:u << 2]"; ++ case 24: ++ return "ld.w %0, %1[%2:t << 2]"; ++ default: ++ internal_error("illegal operand for ldxi"); ++ } ++ } ++ [(set_attr "type" "load") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) ++ ++ ++ ++ ++ ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; sub r8, r7, 8 ++;; st.w r8[0x0], r12 ++;; to ++;; sub r8, r7, 8 ++;; st.w r7[-0x8], r12 ++;;============================================================================= ++; (set (reg:SI 9 r8) ++; (plus:SI (reg/f:SI 6 r7) ++; (const_int ...))) ++; (set (mem:SI (reg:SI 9 r8)) ++; (reg:SI 12 r12)) ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (plus:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" ""))) ++ (set (mem:SI (match_dup 0)) ++ (match_operand:SI 3 "register_operand" ""))] ++ "REGNO(operands[0]) != REGNO(operands[1]) && avr32_const_ok_for_constraint_p(INTVAL(operands[2]), 'K', \"Ks16\")" ++ [(set (match_dup 0) ++ (plus:SI (match_dup 1) ++ (match_dup 2))) ++ (set (mem:SI (plus:SI (match_dup 1) ++ (match_dup 2))) ++ (match_dup 3))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; sub r6, r7, 4 ++;; ld.w r6, r6[0x0] ++;; to ++;; sub r6, r7, 4 ++;; ld.w r6, r7[-0x4] ++;;============================================================================= ++; (set (reg:SI 7 r6) ++; (plus:SI (reg/f:SI 6 r7) ++; (const_int -4 [0xfffffffc]))) ++; (set (reg:SI 7 r6) ++; (mem:SI (reg:SI 7 r6))) ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (plus:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "immediate_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (mem:SI (match_dup 0)))] ++ "REGNO(operands[0]) != REGNO(operands[1]) && avr32_const_ok_for_constraint_p(INTVAL(operands[2]), 'K', \"Ks16\")" ++ [(set (match_dup 0) ++ (plus:SI (match_dup 1) ++ (match_dup 2))) ++ (set (match_dup 3) ++ (mem:SI (plus:SI (match_dup 1) ++ (match_dup 2))))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; ld.sb r0, r7[-0x6] ++;; cashs.b r0 ++;; to ++;; ld.sb r0, r7[-0x6] ++;;============================================================================= ++(define_peephole2 ++ [(set (match_operand:QI 0 "register_operand" "") ++ (match_operand:QI 1 "load_sb_memory_operand" "")) ++ (set (match_operand:SI 2 "register_operand" "") ++ (sign_extend:SI (match_dup 0)))] ++ "(REGNO(operands[0]) == REGNO(operands[2]) || peep2_reg_dead_p(2, operands[0]))" ++ [(set (match_dup 2) ++ (sign_extend:SI (match_dup 1)))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; ld.ub r0, r7[-0x6] ++;; cashu.b r0 ++;; to ++;; ld.ub r0, r7[-0x6] ++;;============================================================================= ++(define_peephole2 ++ [(set (match_operand:QI 0 "register_operand" "") ++ (match_operand:QI 1 "memory_operand" "")) ++ (set (match_operand:SI 2 "register_operand" "") ++ (zero_extend:SI (match_dup 0)))] ++ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 2) ++ (zero_extend:SI (match_dup 1)))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; ld.sh r0, r7[-0x6] ++;; casts.h r0 ++;; to ++;; ld.sh r0, r7[-0x6] ++;;============================================================================= ++(define_peephole2 ++ [(set (match_operand:HI 0 "register_operand" "") ++ (match_operand:HI 1 "memory_operand" "")) ++ (set (match_operand:SI 2 "register_operand" "") ++ (sign_extend:SI (match_dup 0)))] ++ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 2) ++ (sign_extend:SI (match_dup 1)))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; ld.uh r0, r7[-0x6] ++;; castu.h r0 ++;; to ++;; ld.uh r0, r7[-0x6] ++;;============================================================================= ++(define_peephole2 ++ [(set (match_operand:HI 0 "register_operand" "") ++ (match_operand:HI 1 "memory_operand" "")) ++ (set (match_operand:SI 2 "register_operand" "") ++ (zero_extend:SI (match_dup 0)))] ++ "(REGNO(operands[0]) == REGNO(operands[2])) || peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 2) ++ (zero_extend:SI (match_dup 1)))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; mul rd, rx, ry ++;; add rd2, rd ++;; to ++;; mac rd2, rx, ry ++;;============================================================================= ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (mult:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "register_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (plus:SI (match_dup 3) ++ (match_dup 0)))] ++ "peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 3) ++ (plus:SI (mult:SI (match_dup 1) ++ (match_dup 2)) ++ (match_dup 3)))] ++ "") ++ ++ ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Changing ++;; bfextu rd, rs, k5, 1 or and(h/l) rd, one_bit_set_mask ++;; to ++;; bld rs, k5 ++;; ++;; If rd is dead after the operation. ++;;============================================================================= ++(define_peephole2 ++ [ (set (match_operand:SI 0 "register_operand" "") ++ (zero_extract:SI (match_operand:SI 1 "register_operand" "") ++ (const_int 1) ++ (match_operand:SI 2 "immediate_operand" ""))) ++ (set (cc0) ++ (match_dup 0))] ++ "peep2_reg_dead_p(2, operands[0])" ++ [(set (cc0) ++ (and:SI (match_dup 1) ++ (match_dup 2)))] ++ "operands[2] = GEN_INT(1 << INTVAL(operands[2]));") ++ ++(define_peephole2 ++ [ (set (match_operand:SI 0 "register_operand" "") ++ (and:SI (match_operand:SI 1 "register_operand" "") ++ (match_operand:SI 2 "one_bit_set_operand" ""))) ++ (set (cc0) ++ (match_dup 0))] ++ "peep2_reg_dead_p(2, operands[0])" ++ [(set (cc0) ++ (and:SI (match_dup 1) ++ (match_dup 2)))] ++ "") ++ ++;;============================================================================= ++;; Peephole optimizing ++;;----------------------------------------------------------------------------- ++;; Load with extracted index: ld.w Rd, Rb[Ri:{t/u/b/l} << 2] ++;; ++;;============================================================================= ++ ++ ++(define_peephole ++ [(set (match_operand:SI 0 "register_operand" "") ++ (zero_extract:SI (match_operand:SI 1 "register_operand" "") ++ (const_int 8) ++ (match_operand:SI 2 "avr32_extract_shift_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 4 "register_operand" ""))))] ++ ++ "(dead_or_set_p(insn, operands[0]))" ++ { ++ switch ( INTVAL(operands[2]) ){ ++ case 0: ++ return "ld.w %3, %4[%1:b << 2]"; ++ case 8: ++ return "ld.w %3, %4[%1:l << 2]"; ++ case 16: ++ return "ld.w %3, %4[%1:u << 2]"; ++ case 24: ++ return "ld.w %3, %4[%1:t << 2]"; ++ default: ++ internal_error("illegal operand for ldxi"); ++ } ++ } ++ [(set_attr "type" "load") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")] ++ ) ++ ++ ++ ++(define_peephole ++ [(set (match_operand:SI 0 "register_operand" "") ++ (and:SI (match_operand:SI 1 "register_operand" "") (const_int 255))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 3 "register_operand" ""))))] ++ ++ "(dead_or_set_p(insn, operands[0]))" ++ ++ "ld.w %2, %3[%1:b << 2]" ++ [(set_attr "type" "load") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")] ++ ) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (zero_extract:SI (match_operand:SI 1 "register_operand" "") ++ (const_int 8) ++ (match_operand:SI 2 "avr32_extract_shift_operand" ""))) ++ (set (match_operand:SI 3 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 4 "register_operand" ""))))] ++ ++ "(peep2_reg_dead_p(2, operands[0])) ++ || (REGNO(operands[0]) == REGNO(operands[3]))" ++ [(set (match_dup 3) ++ (mem:SI (plus:SI ++ (match_dup 4) ++ (mult:SI (zero_extract:SI (match_dup 1) ++ (const_int 8) ++ (match_dup 2)) ++ (const_int 4)))))] ++ ) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (zero_extend:SI (match_operand:QI 1 "register_operand" ""))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 3 "register_operand" ""))))] ++ ++ "(peep2_reg_dead_p(2, operands[0])) ++ || (REGNO(operands[0]) == REGNO(operands[2]))" ++ [(set (match_dup 2) ++ (mem:SI (plus:SI ++ (match_dup 3) ++ (mult:SI (zero_extract:SI (match_dup 1) ++ (const_int 8) ++ (const_int 0)) ++ (const_int 4)))))] ++ "operands[1] = gen_rtx_REG(SImode, REGNO(operands[1]));" ++ ) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (and:SI (match_operand:SI 1 "register_operand" "") ++ (const_int 255))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 3 "register_operand" ""))))] ++ ++ "(peep2_reg_dead_p(2, operands[0])) ++ || (REGNO(operands[0]) == REGNO(operands[2]))" ++ [(set (match_dup 2) ++ (mem:SI (plus:SI ++ (match_dup 3) ++ (mult:SI (zero_extract:SI (match_dup 1) ++ (const_int 8) ++ (const_int 0)) ++ (const_int 4)))))] ++ "" ++ ) ++ ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (lshiftrt:SI (match_operand:SI 1 "register_operand" "") ++ (const_int 24))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mem:SI (plus:SI (mult:SI (match_dup 0) (const_int 4)) ++ (match_operand:SI 3 "register_operand" ""))))] ++ ++ "(peep2_reg_dead_p(2, operands[0])) ++ || (REGNO(operands[0]) == REGNO(operands[2]))" ++ [(set (match_dup 2) ++ (mem:SI (plus:SI ++ (match_dup 3) ++ (mult:SI (zero_extract:SI (match_dup 1) ++ (const_int 8) ++ (const_int 24)) ++ (const_int 4)))))] ++ "" ++ ) ++ ++ ++;;************************************************ ++;; ANDN ++;; ++;;************************************************ ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (not:SI (match_operand:SI 1 "register_operand" ""))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (and:SI (match_dup 2) ++ (match_dup 0)))] ++ "peep2_reg_dead_p(2, operands[0])" ++ ++ [(set (match_dup 2) ++ (and:SI (match_dup 2) ++ (not:SI (match_dup 1)) ++ ))] ++ "" ++) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (not:SI (match_operand:SI 1 "register_operand" ""))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (and:SI (match_dup 0) ++ (match_dup 2) ++ ))] ++ "peep2_reg_dead_p(2, operands[0])" ++ ++ [(set (match_dup 2) ++ (and:SI (match_dup 2) ++ (not:SI (match_dup 1)) ++ ))] ++ ++ "" ++) ++ ++ ++;;================================================================= ++;; Addabs peephole ++;;================================================================= ++ ++(define_peephole ++ [(set (match_operand:SI 2 "register_operand" "=r") ++ (abs:SI (match_operand:SI 1 "register_operand" "r"))) ++ (set (match_operand:SI 0 "register_operand" "=r") ++ (plus:SI (match_operand:SI 3 "register_operand" "r") ++ (match_dup 2)))] ++ "dead_or_set_p(insn, operands[2])" ++ "addabs %0, %3, %1" ++ [(set_attr "length" "4") ++ (set_attr "cc" "set_z")]) ++ ++(define_peephole ++ [(set (match_operand:SI 2 "register_operand" "=r") ++ (abs:SI (match_operand:SI 1 "register_operand" "r"))) ++ (set (match_operand:SI 0 "register_operand" "=r") ++ (plus:SI (match_dup 2) ++ (match_operand:SI 3 "register_operand" "r")))] ++ "dead_or_set_p(insn, operands[2])" ++ "addabs %0, %3, %1" ++ [(set_attr "length" "4") ++ (set_attr "cc" "set_z")]) ++ ++ ++;;================================================================= ++;; Detect roundings ++;;================================================================= ++ ++(define_insn "*round" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (ashiftrt:SI (plus:SI (match_operand:SI 1 "register_operand" "0") ++ (match_operand:SI 2 "immediate_operand" "i")) ++ (match_operand:SI 3 "immediate_operand" "i")))] ++ "avr32_rnd_operands(operands[2], operands[3])" ++ ++ "satrnds %0 >> %3, 31" ++ ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4")] ++ ++ ) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (plus:SI (match_dup 0) ++ (match_operand:SI 1 "immediate_operand" ""))) ++ (set (match_dup 0) ++ (ashiftrt:SI (match_dup 0) ++ (match_operand:SI 2 "immediate_operand" "")))] ++ "avr32_rnd_operands(operands[1], operands[2])" ++ ++ [(set (match_dup 0) ++ (ashiftrt:SI (plus:SI (match_dup 0) ++ (match_dup 1)) ++ (match_dup 2)))] ++ ) ++ ++(define_peephole ++ [(set (match_operand:SI 0 "register_operand" "r") ++ (plus:SI (match_dup 0) ++ (match_operand:SI 1 "immediate_operand" "i"))) ++ (set (match_dup 0) ++ (ashiftrt:SI (match_dup 0) ++ (match_operand:SI 2 "immediate_operand" "i")))] ++ "avr32_rnd_operands(operands[1], operands[2])" ++ ++ "satrnds %0 >> %2, 31" ++ ++ [(set_attr "type" "alu_sat") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")] ++ ++ ) ++ ++ ++;;================================================================= ++;; mcall ++;;================================================================= ++(define_peephole ++ [(set (match_operand:SI 0 "register_operand" "") ++ (match_operand 1 "avr32_const_pool_ref_operand" "")) ++ (parallel [(call (mem:SI (match_dup 0)) ++ (match_operand 2 "" "")) ++ (clobber (reg:SI LR_REGNUM))])] ++ "dead_or_set_p(insn, operands[0])" ++ "mcall %1" ++ [(set_attr "type" "call") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")] ++) ++ ++(define_peephole ++ [(set (match_operand:SI 2 "register_operand" "") ++ (match_operand 1 "avr32_const_pool_ref_operand" "")) ++ (parallel [(set (match_operand 0 "register_operand" "") ++ (call (mem:SI (match_dup 2)) ++ (match_operand 3 "" ""))) ++ (clobber (reg:SI LR_REGNUM))])] ++ "dead_or_set_p(insn, operands[2])" ++ "mcall %1" ++ [(set_attr "type" "call") ++ (set_attr "length" "4") ++ (set_attr "cc" "call_set")] ++) ++ ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (match_operand 1 "avr32_const_pool_ref_operand" "")) ++ (parallel [(call (mem:SI (match_dup 0)) ++ (match_operand 2 "" "")) ++ (clobber (reg:SI LR_REGNUM))])] ++ "peep2_reg_dead_p(2, operands[0])" ++ [(parallel [(call (mem:SI (match_dup 1)) ++ (match_dup 2)) ++ (clobber (reg:SI LR_REGNUM))])] ++ "" ++) ++ ++(define_peephole2 ++ [(set (match_operand:SI 0 "register_operand" "") ++ (match_operand 1 "avr32_const_pool_ref_operand" "")) ++ (parallel [(set (match_operand 2 "register_operand" "") ++ (call (mem:SI (match_dup 0)) ++ (match_operand 3 "" ""))) ++ (clobber (reg:SI LR_REGNUM))])] ++ "(peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[2]) == REGNO(operands[0])))" ++ [(parallel [(set (match_dup 2) ++ (call (mem:SI (match_dup 1)) ++ (match_dup 3))) ++ (clobber (reg:SI LR_REGNUM))])] ++ "" ++) ++ ++;;================================================================= ++;; Returning a value ++;;================================================================= ++ ++ ++(define_peephole ++ [(set (match_operand 0 "register_operand" "") ++ (match_operand 1 "register_operand" "")) ++ (return)] ++ "USE_RETURN_INSN (TRUE) && (REGNO(operands[0]) == RETVAL_REGNUM) ++ && (REGNO(operands[1]) != LR_REGNUM) ++ && (REGNO_REG_CLASS(REGNO(operands[1])) == GENERAL_REGS)" ++ "retal %1" ++ [(set_attr "type" "call") ++ (set_attr "length" "2")] ++ ) ++ ++ ++(define_peephole ++ [(set (match_operand 0 "register_operand" "r") ++ (match_operand 1 "immediate_operand" "i")) ++ (return)] ++ "(USE_RETURN_INSN (FALSE) && (REGNO(operands[0]) == RETVAL_REGNUM) && ++ ((INTVAL(operands[1]) == -1) || (INTVAL(operands[1]) == 0) || (INTVAL(operands[1]) == 1)))" ++ { ++ avr32_output_return_instruction (TRUE, FALSE, NULL, operands[1]); ++ return ""; ++ } ++ [(set_attr "type" "call") ++ (set_attr "length" "4")] ++ ) ++ ++(define_peephole ++ [(set (match_operand 0 "register_operand" "r") ++ (match_operand 1 "immediate_operand" "i")) ++ (unspec_volatile [(return)] VUNSPEC_EPILOGUE)] ++ "(REGNO(operands[0]) == RETVAL_REGNUM) && ++ ((INTVAL(operands[1]) == -1) || (INTVAL(operands[1]) == 0) || (INTVAL(operands[1]) == 1))" ++ { ++ avr32_output_return_instruction (FALSE, FALSE, NULL, operands[1]); ++ return ""; ++ } ++ ; Length is absolute worst case ++ [(set_attr "type" "branch") ++ (set_attr "length" "12")] ++ ) ++ ++(define_peephole ++ [(set (match_operand 0 "register_operand" "=r") ++ (if_then_else (match_operator 1 "avr32_comparison_operator" ++ [(match_operand 4 "register_operand" "r") ++ (match_operand 5 "register_immediate_operand" "rKs21")]) ++ (match_operand 2 "avr32_cond_register_immediate_operand" "rKs08") ++ (match_operand 3 "avr32_cond_register_immediate_operand" "rKs08"))) ++ (return)] ++ "USE_RETURN_INSN (TRUE) && (REGNO(operands[0]) == RETVAL_REGNUM)" ++ { ++ operands[1] = avr32_output_cmp(operands[1], GET_MODE(operands[4]), operands[4], operands[5]); ++ ++ if ( GET_CODE(operands[2]) == REG ++ && GET_CODE(operands[3]) == REG ++ && REGNO(operands[2]) != LR_REGNUM ++ && REGNO(operands[3]) != LR_REGNUM ){ ++ return "ret%1 %2\;ret%i1 %3"; ++ } else if ( GET_CODE(operands[2]) == REG ++ && GET_CODE(operands[3]) == CONST_INT ){ ++ if ( INTVAL(operands[3]) == -1 ++ || INTVAL(operands[3]) == 0 ++ || INTVAL(operands[3]) == 1 ){ ++ return "ret%1 %2\;ret%i1 %d3"; ++ } else { ++ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; ++ } ++ } else if ( GET_CODE(operands[2]) == CONST_INT ++ && GET_CODE(operands[3]) == REG ){ ++ if ( INTVAL(operands[2]) == -1 ++ || INTVAL(operands[2]) == 0 ++ || INTVAL(operands[2]) == 1 ){ ++ return "ret%1 %d2\;ret%i1 %3"; ++ } else { ++ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; ++ } ++ } else { ++ if ( (INTVAL(operands[2]) == -1 ++ || INTVAL(operands[2]) == 0 ++ || INTVAL(operands[2]) == 1 ) ++ && (INTVAL(operands[3]) == -1 ++ || INTVAL(operands[3]) == 0 ++ || INTVAL(operands[3]) == 1 )){ ++ return "ret%1 %d2\;ret%i1 %d3"; ++ } else { ++ return "mov%1 r12, %2\;mov%i1 r12, %3\;retal r12"; ++ } ++ } ++ } ++ ++ [(set_attr "length" "10") ++ (set_attr "cc" "none") ++ (set_attr "type" "call")]) ++ ++ ++;;================================================================= ++;; mulnhh.w ++;;================================================================= ++ ++(define_peephole2 ++ [(set (match_operand:HI 0 "register_operand" "") ++ (neg:HI (match_operand:HI 1 "register_operand" ""))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mult:SI ++ (sign_extend:SI (match_dup 0)) ++ (sign_extend:SI (match_operand:HI 3 "register_operand" ""))))] ++ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[2]) == REGNO(operands[0]))" ++ [ (set (match_dup 2) ++ (mult:SI ++ (sign_extend:SI (neg:HI (match_dup 1))) ++ (sign_extend:SI (match_dup 3))))] ++ "" ++ ) ++ ++(define_peephole2 ++ [(set (match_operand:HI 0 "register_operand" "") ++ (neg:HI (match_operand:HI 1 "register_operand" ""))) ++ (set (match_operand:SI 2 "register_operand" "") ++ (mult:SI ++ (sign_extend:SI (match_operand:HI 3 "register_operand" "")) ++ (sign_extend:SI (match_dup 0))))] ++ "(peep2_reg_dead_p(2, operands[0])) || (REGNO(operands[2]) == REGNO(operands[0]))" ++ [ (set (match_dup 2) ++ (mult:SI ++ (sign_extend:SI (neg:HI (match_dup 1))) ++ (sign_extend:SI (match_dup 3))))] ++ "" ++ ) ++ ++ ++ ++;;================================================================= ++;; Vector set and extract operations ++;;================================================================= ++(define_insn "vec_setv2hi_hi" ++ [(set (match_operand:V2HI 0 "register_operand" "=r") ++ (vec_merge:V2HI ++ (match_dup 0) ++ (vec_duplicate:V2HI ++ (match_operand:HI 1 "register_operand" "r")) ++ (const_int 1)))] ++ "" ++ "bfins\t%0, %1, 16, 16" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")]) ++ ++(define_insn "vec_setv2hi_lo" ++ [(set (match_operand:V2HI 0 "register_operand" "+r") ++ (vec_merge:V2HI ++ (match_dup 0) ++ (vec_duplicate:V2HI ++ (match_operand:HI 1 "register_operand" "r")) ++ (const_int 2)))] ++ "" ++ "bfins\t%0, %1, 0, 16" ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")]) ++ ++(define_expand "vec_setv2hi" ++ [(set (match_operand:V2HI 0 "register_operand" "") ++ (vec_merge:V2HI ++ (match_dup 0) ++ (vec_duplicate:V2HI ++ (match_operand:HI 1 "register_operand" "")) ++ (match_operand 2 "immediate_operand" "")))] ++ "" ++ { operands[2] = GEN_INT(INTVAL(operands[2]) + 1); } ++ ) ++ ++(define_insn "vec_extractv2hi" ++ [(set (match_operand:HI 0 "register_operand" "=r") ++ (vec_select:HI ++ (match_operand:V2HI 1 "register_operand" "r") ++ (parallel [(match_operand:SI 2 "immediate_operand" "i")])))] ++ "" ++ { ++ if ( INTVAL(operands[2]) == 0 ) ++ return "bfextu\t%0, %1, 16, 16"; ++ else ++ return "bfextu\t%0, %1, 0, 16"; ++ } ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")]) ++ ++(define_insn "vec_extractv4qi" ++ [(set (match_operand:QI 0 "register_operand" "=r") ++ (vec_select:QI ++ (match_operand:V4QI 1 "register_operand" "r") ++ (parallel [(match_operand:SI 2 "immediate_operand" "i")])))] ++ "" ++ { ++ switch ( INTVAL(operands[2]) ){ ++ case 0: ++ return "bfextu\t%0, %1, 24, 8"; ++ case 1: ++ return "bfextu\t%0, %1, 16, 8"; ++ case 2: ++ return "bfextu\t%0, %1, 8, 8"; ++ case 3: ++ return "bfextu\t%0, %1, 0, 8"; ++ default: ++ abort(); ++ } ++ } ++ [(set_attr "type" "alu") ++ (set_attr "length" "4") ++ (set_attr "cc" "clobber")]) ++ ++ ++(define_insn "concatv2hi" ++ [(set (match_operand:V2HI 0 "register_operand" "=r, r, r") ++ (vec_concat:V2HI ++ (match_operand:HI 1 "register_operand" "r, r, 0") ++ (match_operand:HI 2 "register_operand" "r, 0, r")))] ++ "" ++ "@ ++ mov\t%0, %1\;bfins\t%0, %2, 0, 16 ++ bfins\t%0, %2, 0, 16 ++ bfins\t%0, %1, 16, 16" ++ [(set_attr "length" "6, 4, 4") ++ (set_attr "type" "alu")]) ++ ++ ++;; Load the atomic operation description ++(include "sync.md") ++ ++;; Load the SIMD description ++(include "simd.md") ++ ++;; Load the FP coprocessor patterns ++(include "fpcp.md") +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-modes.def gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-modes.def +--- gcc-4.2.1/gcc/config/avr32/avr32-modes.def 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-modes.def 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1 @@ ++VECTOR_MODES (INT, 4); /* V4QI V2HI */ +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.opt gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.opt +--- gcc-4.2.1/gcc/config/avr32/avr32.opt 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32.opt 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,73 @@ ++; Options for the ATMEL AVR32 port of the compiler. ++ ++; Copyright 2007 Atmel Corporation. ++; ++; This file is part of GCC. ++; ++; GCC is free software; you can redistribute it and/or modify it under ++; the terms of the GNU General Public License as published by the Free ++; Software Foundation; either version 2, or (at your option) any later ++; version. ++; ++; GCC is distributed in the hope that it will be useful, but WITHOUT ANY ++; WARRANTY; without even the implied warranty of MERCHANTABILITY or ++; FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License ++; for more details. ++; ++; You should have received a copy of the GNU General Public License ++; along with GCC; see the file COPYING. If not, write to the Free ++; Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA ++; 02110-1301, USA. ++ ++muse-rodata-section ++Target Report Mask(USE_RODATA_SECTION) ++Do not put readonly-data in .text section, but in .rodata. ++ ++mhard-float ++Target Report Undocumented Mask(HARD_FLOAT) ++Use floating point coprocessor instructions. ++ ++msoft-float ++Target Report Undocumented InverseMask(HARD_FLOAT, SOFT_FLOAT) ++Use software floating-point library for floating-point operations. ++ ++force-double-align ++Target Report RejectNegative Mask(FORCE_DOUBLE_ALIGN) ++Force double-word alignment for double-word memory accesses. ++ ++mno-init-got ++Target Report RejectNegative Mask(NO_INIT_GOT) ++Do not initialize GOT register before using it when compiling PIC code. ++ ++mrelax ++Target Report Mask(RELAX) ++Let invoked assembler and linker do relaxing (Enabled by default when optimization level is >1). ++ ++mmd-reorg-opt ++Target Report Undocumented Mask(MD_REORG_OPTIMIZATION) ++Perform machine dependent optimizations in reorg stage. ++ ++masm-addr-pseudos ++Target Report Mask(HAS_ASM_ADDR_PSEUDOS) ++Use assembler pseudo-instructions lda.w and call for handling direct addresses. (Enabled by default) ++ ++mpart= ++Target Report RejectNegative Joined Var(avr32_part_name) ++Specify the AVR32 part name ++ ++mcpu= ++Target Report RejectNegative Joined Undocumented Var(avr32_part_name) ++Specify the AVR32 part name (deprecated) ++ ++march= ++Target Report RejectNegative Joined Var(avr32_arch_name) ++Specify the AVR32 architecture name ++ ++mfast-float ++Target Report Mask(FAST_FLOAT) ++Enable fast floating-point library. Enabled by default if the -funsafe-math-optimizations switch is specified. ++ ++mimm-in-const-pool ++Target Report Var(avr32_imm_in_const_pool) Init(-1) ++Put large immediates in constant pool. This is enabled by default for archs with insn-cache. ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32-protos.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-protos.h +--- gcc-4.2.1/gcc/config/avr32/avr32-protos.h 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/avr32-protos.h 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,185 @@ ++/* ++ Prototypes for exported functions defined in avr32.c ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ Initial porting by Anders �dland. ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++ ++#ifndef AVR32_PROTOS_H ++#define AVR32_PROTOS_H ++ ++extern const int swap_reg[]; ++ ++extern int avr32_valid_macmac_bypass (rtx, rtx); ++extern int avr32_valid_mulmac_bypass (rtx, rtx); ++ ++extern int avr32_decode_lcomm_symbol_offset (rtx, int *); ++extern void avr32_encode_lcomm_symbol_offset (tree, char *, int); ++ ++extern const char *avr32_strip_name_encoding (const char *); ++ ++extern rtx avr32_get_note_reg_equiv (rtx insn); ++ ++extern int avr32_use_return_insn (int iscond); ++ ++extern void avr32_make_reglist16 (int reglist16_vect, char *reglist16_string); ++ ++extern void avr32_make_reglist8 (int reglist8_vect, char *reglist8_string); ++extern void avr32_make_fp_reglist_w (int reglist_mask, char *reglist_string); ++extern void avr32_make_fp_reglist_d (int reglist_mask, char *reglist_string); ++ ++extern void avr32_output_return_instruction (int single_ret_inst, ++ int iscond, rtx cond, ++ rtx r12_imm); ++extern void avr32_expand_prologue (void); ++extern void avr32_set_return_address (rtx source, rtx scratch); ++ ++extern int avr32_hard_regno_mode_ok (int regno, enum machine_mode mode); ++extern int avr32_extra_constraint_s (rtx value, const int strict); ++extern int avr32_eh_return_data_regno (const int n); ++extern int avr32_initial_elimination_offset (const int from, const int to); ++extern rtx avr32_function_arg (CUMULATIVE_ARGS * cum, enum machine_mode mode, ++ tree type, int named); ++extern void avr32_init_cumulative_args (CUMULATIVE_ARGS * cum, tree fntype, ++ rtx libname, tree fndecl); ++extern void avr32_function_arg_advance (CUMULATIVE_ARGS * cum, ++ enum machine_mode mode, ++ tree type, int named); ++#ifdef ARGS_SIZE_RTX ++/* expr.h defines ARGS_SIZE_RTX and `enum direction'. */ ++extern enum direction avr32_function_arg_padding (enum machine_mode mode, ++ tree type); ++#endif /* ARGS_SIZE_RTX */ ++extern rtx avr32_function_value (tree valtype, tree func, bool outgoing); ++extern rtx avr32_libcall_value (enum machine_mode mode); ++extern int avr32_sched_use_dfa_pipeline_interface (void); ++extern bool avr32_return_in_memory (tree type, tree fntype); ++extern void avr32_regs_to_save (char *operand); ++extern void avr32_target_asm_function_prologue (FILE * file, ++ HOST_WIDE_INT size); ++extern void avr32_target_asm_function_epilogue (FILE * file, ++ HOST_WIDE_INT size); ++extern void avr32_trampoline_template (FILE * file); ++extern void avr32_initialize_trampoline (rtx addr, rtx fnaddr, ++ rtx static_chain); ++extern int avr32_legitimate_address (enum machine_mode mode, rtx x, ++ int strict); ++extern int avr32_legitimate_constant_p (rtx x); ++ ++extern int avr32_legitimate_pic_operand_p (rtx x); ++ ++extern rtx avr32_find_symbol (rtx x); ++extern void avr32_select_section (rtx exp, int reloc, int align); ++extern void avr32_encode_section_info (tree decl, rtx rtl, int first); ++extern void avr32_asm_file_end (FILE * stream); ++extern void avr32_asm_output_ascii (FILE * stream, char *ptr, int len); ++extern void avr32_asm_output_common (FILE * stream, const char *name, ++ int size, int rounded); ++extern void avr32_asm_output_label (FILE * stream, const char *name); ++extern void avr32_asm_declare_object_name (FILE * stream, char *name, ++ tree decl); ++extern void avr32_asm_globalize_label (FILE * stream, const char *name); ++extern void avr32_asm_weaken_label (FILE * stream, const char *name); ++extern void avr32_asm_output_external (FILE * stream, tree decl, ++ const char *name); ++extern void avr32_asm_output_external_libcall (FILE * stream, rtx symref); ++extern void avr32_asm_output_labelref (FILE * stream, const char *name); ++extern void avr32_notice_update_cc (rtx exp, rtx insn); ++extern void avr32_print_operand (FILE * stream, rtx x, int code); ++extern void avr32_print_operand_address (FILE * stream, rtx x); ++ ++extern int avr32_symbol (rtx x); ++ ++extern void avr32_select_rtx_section (enum machine_mode mode, rtx x, ++ unsigned HOST_WIDE_INT align); ++ ++extern int avr32_load_multiple_operation (rtx op, enum machine_mode mode); ++extern int avr32_store_multiple_operation (rtx op, enum machine_mode mode); ++ ++extern int avr32_const_ok_for_constraint_p (HOST_WIDE_INT value, char c, ++ const char *str); ++ ++extern bool avr32_cannot_force_const_mem (rtx x); ++ ++extern void avr32_init_builtins (void); ++ ++extern rtx avr32_expand_builtin (tree exp, rtx target, rtx subtarget, ++ enum machine_mode mode, int ignore); ++ ++extern bool avr32_must_pass_in_stack (enum machine_mode mode, tree type); ++ ++extern bool avr32_strict_argument_naming (CUMULATIVE_ARGS * ca); ++ ++extern bool avr32_pass_by_reference (CUMULATIVE_ARGS * cum, ++ enum machine_mode mode, ++ tree type, bool named); ++ ++extern rtx avr32_gen_load_multiple (rtx * regs, int count, rtx from, ++ int write_back, int in_struct_p, ++ int scalar_p); ++extern rtx avr32_gen_store_multiple (rtx * regs, int count, rtx to, ++ int in_struct_p, int scalar_p); ++extern int avr32_gen_movmemsi (rtx * operands); ++ ++extern int avr32_rnd_operands (rtx add, rtx shift); ++extern int avr32_adjust_insn_length (rtx insn, int length); ++ ++extern int symbol_mentioned_p (rtx x); ++extern int label_mentioned_p (rtx x); ++extern rtx legitimize_pic_address (rtx orig, enum machine_mode mode, rtx reg); ++extern int avr32_address_register_rtx_p (rtx x, int strict_p); ++extern int avr32_legitimate_index_p (enum machine_mode mode, rtx index, ++ int strict_p); ++ ++extern int avr32_const_double_immediate (rtx value); ++extern void avr32_init_expanders (void); ++extern rtx avr32_return_addr (int count, rtx frame); ++extern bool avr32_got_mentioned_p (rtx addr); ++ ++extern void avr32_final_prescan_insn (rtx insn, rtx * opvec, int noperands); ++ ++extern int avr32_expand_movcc (enum machine_mode mode, rtx operands[]); ++extern int avr32_expand_addcc (enum machine_mode mode, rtx operands[]); ++#ifdef RTX_CODE ++extern int avr32_expand_scc (RTX_CODE cond, rtx * operands); ++#endif ++ ++extern int avr32_store_bypass (rtx insn_out, rtx insn_in); ++extern int avr32_mul_waw_bypass (rtx insn_out, rtx insn_in); ++extern int avr32_valid_load_double_bypass (rtx insn_out, rtx insn_in); ++extern int avr32_valid_load_quad_bypass (rtx insn_out, rtx insn_in); ++extern rtx avr32_output_cmp (rtx cond, enum machine_mode mode, ++ rtx op0, rtx op1); ++ ++rtx get_next_insn_cond (rtx cur_insn); ++int set_next_insn_cond (rtx cur_insn, rtx cond); ++void avr32_override_options (void); ++void avr32_load_pic_register (void); ++void avr32_optimization_options (int level, int size); ++void avr32_split_const_expr (enum machine_mode mode, ++ enum machine_mode new_mode, ++ rtx expr, ++ rtx *split_expr); ++void avr32_get_intval (enum machine_mode mode, ++ rtx const_expr, ++ HOST_WIDE_INT *val); ++ ++ ++#endif /* AVR32_PROTOS_H */ +diff -Nrup gcc-4.2.1/gcc/config/avr32/crti.asm gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crti.asm +--- gcc-4.2.1/gcc/config/avr32/crti.asm 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crti.asm 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,64 @@ ++/* ++ Init/fini stuff for AVR32. ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++ ++/* The code in sections .init and .fini is supposed to be a single ++ regular function. The function in .init is called directly from ++ start in crt1.asm. The function in .fini is atexit()ed in crt1.asm ++ too. ++ ++ crti.asm contributes the prologue of a function to these sections, ++ and crtn.asm comes up the epilogue. STARTFILE_SPEC should list ++ crti.o before any other object files that might add code to .init ++ or .fini sections, and ENDFILE_SPEC should list crtn.o after any ++ such object files. */ ++ ++ .file "crti.asm" ++ ++ .section ".init" ++/* Just load the GOT */ ++ .align 2 ++ .global _init ++_init: ++ stm --sp, r6, lr ++ lddpc r6, 1f ++0: ++ rsub r6, pc ++ rjmp 2f ++ .align 2 ++1: .long 0b - _GLOBAL_OFFSET_TABLE_ ++2: ++ ++ .section ".fini" ++/* Just load the GOT */ ++ .align 2 ++ .global _fini ++_fini: ++ stm --sp, r6, lr ++ lddpc r6, 1f ++0: ++ rsub r6, pc ++ rjmp 2f ++ .align 2 ++1: .long 0b - _GLOBAL_OFFSET_TABLE_ ++2: ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/crtn.asm gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crtn.asm +--- gcc-4.2.1/gcc/config/avr32/crtn.asm 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/crtn.asm 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,44 @@ ++/* Copyright (C) 2001 Free Software Foundation, Inc. ++ Written By Nick Clifton ++ ++ This file is free software; you can redistribute it and/or modify it ++ under the terms of the GNU General Public License as published by the ++ Free Software Foundation; either version 2, or (at your option) any ++ later version. ++ ++ In addition to the permissions in the GNU General Public License, the ++ Free Software Foundation gives you unlimited permission to link the ++ compiled version of this file with other programs, and to distribute ++ those programs without any restriction coming from the use of this ++ file. (The General Public License restrictions do apply in other ++ respects; for example, they cover modification of the file, and ++ distribution when not linked into another program.) ++ ++ This file is distributed in the hope that it will be useful, but ++ WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ++ General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; see the file COPYING. If not, write to ++ the Free Software Foundation, 59 Temple Place - Suite 330, ++ Boston, MA 02111-1307, USA. ++ ++ As a special exception, if you link this library with files ++ compiled with GCC to produce an executable, this does not cause ++ the resulting executable to be covered by the GNU General Public License. ++ This exception does not however invalidate any other reasons why ++ the executable file might be covered by the GNU General Public License. ++*/ ++ ++ ++ ++ ++ .file "crtn.asm" ++ ++ .section ".init" ++ ldm sp++, r6, pc ++ ++ .section ".fini" ++ ldm sp++, r6, pc ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/fpcp.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/fpcp.md +--- gcc-4.2.1/gcc/config/avr32/fpcp.md 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/fpcp.md 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,551 @@ ++;; AVR32 machine description file for Floating-Point instructions. ++;; Copyright 2003-2006 Atmel Corporation. ++;; ++;; Written by Ronny Pedersen, Atmel Norway, ++;; ++;; This file is part of GCC. ++;; ++;; This program is free software; you can redistribute it and/or modify ++;; it under the terms of the GNU General Public License as published by ++;; the Free Software Foundation; either version 2 of the License, or ++;; (at your option) any later version. ++;; ++;; This program is distributed in the hope that it will be useful, ++;; but WITHOUT ANY WARRANTY; without even the implied warranty of ++;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++;; GNU General Public License for more details. ++;; ++;; You should have received a copy of the GNU General Public License ++;; along with this program; if not, write to the Free Software ++;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. ++ ++;; -*- Mode: Scheme -*- ++ ++;;****************************************************************************** ++;; Automaton pipeline description for floating-point coprocessor insns ++;;****************************************************************************** ++(define_cpu_unit "fid,fm1,fm2,fm3,fm4,fwb,fcmp,fcast" "avr32_ap") ++ ++(define_insn_reservation "fmv_op" 1 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fmv")) ++ "is,da,d,fid,fwb") ++ ++(define_insn_reservation "fmul_op" 5 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fmul")) ++ "is,da,d,fid,fm1,fm2,fm3,fm4,fwb") ++ ++(define_insn_reservation "fcmps_op" 1 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fcmps")) ++ "is,da,d,fid,fcmp") ++ ++(define_insn_reservation "fcmpd_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fcmpd")) ++ "is,da,d,fid*2,fcmp") ++ ++(define_insn_reservation "fcast_op" 3 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fcast")) ++ "is,da,d,fid,fcmp,fcast,fwb") ++ ++(define_insn_reservation "fmvcpu_op" 2 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fmvcpu")) ++ "is,da,d") ++ ++(define_insn_reservation "fldd_op" 1 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fldd")) ++ "is,da,d,fwb") ++ ++(define_insn_reservation "flds_op" 1 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "flds")) ++ "is,da,d,fwb") ++ ++(define_insn_reservation "fsts_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fsts")) ++ "is,da*2,d") ++ ++(define_insn_reservation "fstd_op" 0 ++ (and (eq_attr "pipeline" "ap") ++ (eq_attr "type" "fstd")) ++ "is,da*2,d") ++ ++ ++(define_insn "*movsf_fpcp" ++ [(set (match_operand:SF 0 "nonimmediate_operand" "=f,f,r,f,m,r,r,r,m") ++ (match_operand:SF 1 "general_operand" " f,r,f,m,f,r,G,m,r"))] ++ "TARGET_HARD_FLOAT" ++ "@ ++ fmov.s\t%0, %1 ++ fmov.s\t%0, %1 ++ fmov.s\t%0, %1 ++ fld.s\t%0, %1 ++ fst.s\t%0, %1 ++ mov\t%0, %1 ++ mov\t%0, %1 ++ ld.w\t%0, %1 ++ st.w\t%0, %1" ++ [(set_attr "length" "4,4,4,4,4,2,4,4,4") ++ (set_attr "type" "fmv,flds,fmvcpu,flds,fsts,alu,alu,load,store")]) ++ ++(define_insn_and_split "*movdf_fpcp" ++ [(set (match_operand:DF 0 "nonimmediate_operand" "=f,f,r,f,m,r,r,m") ++ (match_operand:DF 1 "general_operand" " f,r,f,m,f,r,m,r"))] ++ "TARGET_HARD_FLOAT" ++ "@ ++ fmov.d\t%0, %1 ++ fmov.d\t%0, %1 ++ fmov.d\t%0, %1 ++ fld.d\t%0, %1 ++ fst.d\t%0, %1 ++ mov\t%0, %1\;mov\t%m0, %m1 ++ ld.d\t%0, %1 ++ st.d\t%0, %1" ++ ++ "TARGET_HARD_FLOAT ++ && reload_completed ++ && (REG_P(operands[0]) && (REGNO_REG_CLASS(REGNO(operands[0])) == GENERAL_REGS)) ++ && (REG_P(operands[1]) && (REGNO_REG_CLASS(REGNO(operands[1])) == GENERAL_REGS))" ++ [(set (match_dup 0) (match_dup 1)) ++ (set (match_dup 2) (match_dup 3))] ++ " ++ { ++ operands[2] = gen_highpart (SImode, operands[0]); ++ operands[0] = gen_lowpart (SImode, operands[0]); ++ operands[3] = gen_highpart(SImode, operands[1]); ++ operands[1] = gen_lowpart(SImode, operands[1]); ++ } ++ " ++ ++ [(set_attr "length" "4,4,4,4,4,4,4,4") ++ (set_attr "type" "fmv,fldd,fmvcpu,fldd,fstd,alu2,load2,store2")]) ++ ++ ++(define_insn "mulsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fmul.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "nmulsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f"))))] ++ "TARGET_HARD_FLOAT" ++ "fnmul.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") ++ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") ++ (match_operand:SF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:SF 3 "avr32_fp_register_operand" "") ++ (neg:SF (match_dup 0)))] ++ "TARGET_HARD_FLOAT && ++ (peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[3]) == REGNO(operands[0])))" ++ [(set (match_dup 3) ++ (neg:SF (mult:SF (match_dup 1) ++ (match_dup 2))))] ++) ++ ++ ++(define_insn "macsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (plus:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f")) ++ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fmac.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "nmacsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (plus:SF (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f"))) ++ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fnmac.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") ++ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") ++ (match_operand:SF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:SF 3 "avr32_fp_register_operand" "") ++ (minus:SF ++ (match_dup 3) ++ (match_dup 0)))] ++ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 3) ++ (plus:SF (neg:SF (mult:SF (match_dup 1) ++ (match_dup 2))) ++ (match_dup 3)))] ++) ++ ++ ++(define_insn "msubacsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (minus:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f")) ++ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fmsc.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "") ++ (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "") ++ (match_operand:SF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:SF 3 "avr32_fp_register_operand" "") ++ (minus:SF ++ (match_dup 0) ++ (match_dup 3)))] ++ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 3) ++ (minus:SF (mult:SF (match_dup 1) ++ (match_dup 2)) ++ (match_dup 3)))] ++) ++ ++(define_insn "nmsubacsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (minus:SF (neg:SF (mult:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f"))) ++ (match_operand:SF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fnmsc.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++ ++ ++(define_insn "addsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (plus:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fadd.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "subsf3" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (minus:SF (match_operand:SF 1 "avr32_fp_register_operand" "f") ++ (match_operand:SF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fsub.s\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++ ++(define_insn "negsf2" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (neg:SF (match_operand:SF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fneg.s\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmv")]) ++ ++(define_insn "abssf2" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (abs:SF (match_operand:SF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fabs.s\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmv")]) ++ ++(define_insn "truncdfsf2" ++ [(set (match_operand:SF 0 "avr32_fp_register_operand" "=f") ++ (float_truncate:SF ++ (match_operand:DF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fcastd.s\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fcast")]) ++ ++(define_insn "extendsfdf2" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (float_extend:DF ++ (match_operand:SF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fcasts.d\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fcast")]) ++ ++(define_insn "muldf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fmul.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "nmuldf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f"))))] ++ "TARGET_HARD_FLOAT" ++ "fnmul.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") ++ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") ++ (match_operand:DF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:DF 3 "avr32_fp_register_operand" "") ++ (neg:DF (match_dup 0)))] ++ "TARGET_HARD_FLOAT && ++ (peep2_reg_dead_p(2, operands[0]) || (REGNO(operands[3]) == REGNO(operands[0])))" ++ [(set (match_dup 3) ++ (neg:DF (mult:DF (match_dup 1) ++ (match_dup 2))))] ++) ++ ++(define_insn "macdf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (plus:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f")) ++ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fmac.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "msubacdf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (minus:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f")) ++ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fmsc.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") ++ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") ++ (match_operand:DF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:DF 3 "avr32_fp_register_operand" "") ++ (minus:DF ++ (match_dup 0) ++ (match_dup 3)))] ++ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 3) ++ (minus:DF (mult:DF (match_dup 1) ++ (match_dup 2)) ++ (match_dup 3)))] ++ ) ++ ++(define_insn "nmsubacdf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (minus:DF (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f"))) ++ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fnmsc.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "nmacdf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (plus:DF (neg:DF (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f"))) ++ (match_operand:DF 3 "avr32_fp_register_operand" "0")))] ++ "TARGET_HARD_FLOAT" ++ "fnmac.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_peephole2 ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "") ++ (mult:DF (match_operand:DF 1 "avr32_fp_register_operand" "") ++ (match_operand:DF 2 "avr32_fp_register_operand" ""))) ++ (set (match_operand:DF 3 "avr32_fp_register_operand" "") ++ (minus:DF ++ (match_dup 3) ++ (match_dup 0)))] ++ "TARGET_HARD_FLOAT && peep2_reg_dead_p(2, operands[0])" ++ [(set (match_dup 3) ++ (plus:DF (neg:DF (mult:DF (match_dup 1) ++ (match_dup 2))) ++ (match_dup 3)))] ++) ++ ++(define_insn "adddf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (plus:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fadd.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "subdf3" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (minus:DF (match_operand:DF 1 "avr32_fp_register_operand" "f") ++ (match_operand:DF 2 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fsub.d\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmul")]) ++ ++(define_insn "negdf2" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (neg:DF (match_operand:DF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fneg.d\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmv")]) ++ ++(define_insn "absdf2" ++ [(set (match_operand:DF 0 "avr32_fp_register_operand" "=f") ++ (abs:DF (match_operand:DF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ "fabs.d\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmv")]) ++ ++ ++(define_expand "cmpdf" ++ [(set (cc0) ++ (compare:DF ++ (match_operand:DF 0 "general_operand" "") ++ (match_operand:DF 1 "general_operand" "")))] ++ "TARGET_HARD_FLOAT" ++ "{ ++ rtx tmpreg; ++ if ( !REG_P(operands[0]) ) ++ operands[0] = force_reg(DFmode, operands[0]); ++ ++ if ( !REG_P(operands[1]) ) ++ operands[1] = force_reg(DFmode, operands[1]); ++ ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ ++ emit_insn(gen_cmpdf_internal(operands[0], operands[1])); ++ ++ tmpreg = gen_reg_rtx(SImode); ++ emit_insn(gen_fpcc_to_reg(tmpreg)); ++ emit_insn(gen_reg_to_cc(tmpreg)); ++ ++ DONE; ++ }" ++) ++ ++(define_insn "cmpdf_internal" ++ [(set (reg:CC FPCC_REGNUM) ++ (compare:CC ++ (match_operand:DF 0 "avr32_fp_register_operand" "f") ++ (match_operand:DF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ { ++ if (!rtx_equal_p(cc_prev_status.mdep.fpvalue, SET_SRC(PATTERN (insn))) ) ++ return "fcmp.d\t%0, %1"; ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "type" "fcmpd") ++ (set_attr "cc" "fpcompare")]) ++ ++(define_expand "cmpsf" ++ [(set (cc0) ++ (compare:SF ++ (match_operand:SF 0 "general_operand" "") ++ (match_operand:SF 1 "general_operand" "")))] ++ "TARGET_HARD_FLOAT" ++ "{ ++ rtx tmpreg; ++ if ( !REG_P(operands[0]) ) ++ operands[0] = force_reg(SFmode, operands[0]); ++ ++ if ( !REG_P(operands[1]) ) ++ operands[1] = force_reg(SFmode, operands[1]); ++ ++ avr32_compare_op0 = operands[0]; ++ avr32_compare_op1 = operands[1]; ++ ++ emit_insn(gen_cmpsf_internal(operands[0], operands[1])); ++ ++ tmpreg = gen_reg_rtx(SImode); ++ emit_insn(gen_fpcc_to_reg(tmpreg)); ++ emit_insn(gen_reg_to_cc(tmpreg)); ++ ++ DONE; ++ }" ++) ++ ++(define_insn "cmpsf_internal" ++ [(set (reg:CC FPCC_REGNUM) ++ (compare:CC ++ (match_operand:SF 0 "avr32_fp_register_operand" "f") ++ (match_operand:SF 1 "avr32_fp_register_operand" "f")))] ++ "TARGET_HARD_FLOAT" ++ { ++ if (!rtx_equal_p(cc_prev_status.mdep.fpvalue, SET_SRC(PATTERN (insn))) ) ++ return "fcmp.s\t%0, %1"; ++ return ""; ++ } ++ [(set_attr "length" "4") ++ (set_attr "type" "fcmps") ++ (set_attr "cc" "fpcompare")]) ++ ++(define_insn "fpcc_to_reg" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (unspec:SI [(reg:CC FPCC_REGNUM)] ++ UNSPEC_FPCC_TO_REG))] ++ "TARGET_HARD_FLOAT" ++ "fmov.s\t%0, fsr" ++ [(set_attr "length" "4") ++ (set_attr "type" "fmvcpu")]) ++ ++(define_insn "reg_to_cc" ++ [(set (cc0) ++ (unspec:SI [(match_operand:SI 0 "register_operand" "r")] ++ UNSPEC_REG_TO_CC))] ++ "TARGET_HARD_FLOAT" ++ "musfr\t%0" ++ [(set_attr "length" "2") ++ (set_attr "type" "alu") ++ (set_attr "cc" "from_fpcc")]) ++ ++(define_insn "stm_fp" ++ [(unspec [(match_operand 0 "register_operand" "r") ++ (match_operand 1 "const_int_operand" "") ++ (match_operand 2 "const_int_operand" "")] ++ UNSPEC_STMFP)] ++ "TARGET_HARD_FLOAT" ++ { ++ int cop_reglist = INTVAL(operands[1]); ++ ++ if (INTVAL(operands[2]) != 0) ++ return "stcm.w\tcp0, --%0, %C1"; ++ else ++ return "stcm.w\tcp0, %0, %C1"; ++ ++ if ( cop_reglist & ~0xff ){ ++ operands[1] = GEN_INT(cop_reglist & ~0xff); ++ if (INTVAL(operands[2]) != 0) ++ return "stcm.d\tcp0, --%0, %D1"; ++ else ++ return "stcm.d\tcp0, %0, %D1"; ++ } ++ } ++ [(set_attr "type" "fstm") ++ (set_attr "length" "4") ++ (set_attr "cc" "none")]) +diff -Nrup gcc-4.2.1/gcc/config/avr32/lib1funcs.S gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib1funcs.S +--- gcc-4.2.1/gcc/config/avr32/lib1funcs.S 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib1funcs.S 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,2589 @@ ++ ++/* Adjust the unpacked double number if it is a subnormal number. ++ The exponent and mantissa pair are stored ++ in [mant_hi,mant_lo] and [exp]. A register with the correct sign bit in ++ the MSB is passed in [sign]. Needs two scratch ++ registers [scratch1] and [scratch2]. An adjusted and packed double float ++ is present in [mant_hi,mant_lo] after macro has executed */ ++.macro adjust_subnormal_df exp, mant_lo, mant_hi, sign, scratch1, scratch2 ++ /* We have an exponent which is <=0 indicating a subnormal number ++ As it should be stored as if the exponent was 1 (although the ++ exponent field is all zeros to indicate a subnormal number) ++ we have to shift down the mantissa to its correct position. */ ++ neg \exp ++ sub \exp,-1 /* amount to shift down */ ++ cp.w \exp,54 ++ brlo 50f /* if more than 53 shift steps, the ++ entire mantissa will disappear ++ without any rounding to occur */ ++ mov \mant_hi, 0 ++ mov \mant_lo, 0 ++ rjmp 52f ++50: ++ sub \exp,-10 /* do the shift to position the ++ mantissa at the same time ++ note! this does not include the ++ final 1 step shift to add the sign */ ++ ++ /* when shifting, save all shifted out bits in [scratch2]. we may need to ++ look at them to make correct rounding. */ ++ ++ rsub \scratch1,\exp,32 /* get inverted shift count */ ++ cp.w \exp,32 /* handle shifts >= 32 separately */ ++ brhs 51f ++ ++ /* small (<32) shift amount, both words are part of the shift */ ++ lsl \scratch2,\mant_lo,\scratch1 /* save bits to shift out from lsw*/ ++ lsl \scratch1,\mant_hi,\scratch1 /* get bits from msw destined for lsw*/ ++ lsr \mant_lo,\mant_lo,\exp /* shift down lsw */ ++ lsr \mant_hi,\mant_hi,\exp /* shift down msw */ ++ or \mant_hi,\scratch1 /* add bits from msw with prepared lsw */ ++ rjmp 50f ++ ++ /* large (>=32) shift amount, only lsw will have bits left after shift. ++ note that shift operations will use ((shift count) mod 32) so ++ we do not need to subtract 32 from shift count. */ ++51: ++ lsl \scratch2,\mant_hi,\scratch1 /* save bits to shift out from msw */ ++ or \scratch2,\mant_lo /* also save all bits from lsw */ ++ mov \mant_lo,\mant_hi /* msw -> lsw (i.e. "shift 32 first") */ ++ mov \mant_hi,0 /* clear msw */ ++ lsr \mant_lo,\mant_lo,\exp /* make rest of shift inside lsw */ ++ ++50: ++ /* result is almost ready to return, except that least significant bit ++ and the part we already shifted out may cause the result to be ++ rounded */ ++ bld \mant_lo,0 /* get bit to be shifted out */ ++ brcc 51f /* if bit was 0, no rounding */ ++ ++ /* msb of part to remove is 1, so rounding depends on rest of bits */ ++ tst \scratch2,\scratch2 /* get shifted out tail */ ++ brne 50f /* if rest > 0, do round */ ++ bld \mant_lo,1 /* we have to look at lsb in result */ ++ brcc 51f /* if lsb is 0, don't round */ ++ ++50: ++ /* subnormal result requires rounding ++ rounding may cause subnormal to become smallest normal number ++ luckily, smallest normal number has exactly the representation ++ we got by rippling a one bit up from mantissa into exponent field. */ ++ sub \mant_lo,-1 ++ subcc \mant_hi,-1 ++ ++51: ++ /* shift and return packed double with correct sign */ ++ rol \sign ++ ror \mant_hi ++ ror \mant_lo ++52: ++.endm ++ ++ ++/* Adjust subnormal single float number with exponent [exp] ++ and mantissa [mant] and round. */ ++.macro adjust_subnormal_sf sf, exp, mant, sign, scratch ++ /* subnormal number */ ++ rsub \exp,\exp, 1 /* shift amount */ ++ cp.w \exp, 25 ++ movhs \mant, 0 ++ brhs 90f /* Return zero */ ++ rsub \scratch, \exp, 32 ++ lsl \scratch, \mant,\scratch/* Check if there are any bits set ++ in the bits discarded in the mantissa */ ++ srne \scratch /* If so set the lsb of the shifted mantissa */ ++ lsr \mant,\mant,\exp /* Shift the mantissa */ ++ or \mant, \scratch /* Round lsb if any bits were shifted out */ ++ /* Rounding : For explaination, see round_sf. */ ++ mov \scratch, 0x7f /* Set rounding constant */ ++ bld \mant, 8 ++ subeq \scratch, -1 /* For odd numbers use rounding constant 0x80 */ ++ add \mant, \scratch /* Add rounding constant to mantissa */ ++ /* We can't overflow because mantissa is at least shifted one position ++ to the right so the implicit bit is zero. We can however get the implicit ++ bit set after rounding which means that we have the lowest normal number ++ but this is ok since this bit has the same position as the LSB of the ++ exponent */ ++ lsr \sf, \mant, 7 ++ /* Rotate in sign */ ++ lsl \sign, 1 ++ ror \sf ++90: ++.endm ++ ++ ++/* Round the unpacked df number with exponent [exp] and ++ mantissa [mant_hi, mant_lo]. Uses scratch register ++ [scratch] */ ++.macro round_df exp, mant_lo, mant_hi, scratch ++ mov \scratch, 0x3ff /* Rounding constant */ ++ bld \mant_lo,11 /* Check if lsb in the final result is ++ set */ ++ subeq \scratch, -1 /* Adjust rounding constant to 0x400 ++ if rounding 0.5 upwards */ ++ add \mant_lo, \scratch /* Round */ ++ acr \mant_hi /* If overflowing we know that ++ we have all zeros in the bits not ++ scaled out so we can leave them ++ but we must increase the exponent with ++ two since we had an implicit bit ++ which is lost + the extra overflow bit */ ++ subcs \exp, -2 /* Update exponent */ ++.endm ++ ++/* Round single float number stored in [mant] and [exp] */ ++.macro round_sf exp, mant, scratch ++ /* Round: ++ For 0.5 we round to nearest even integer ++ for all other cases we round to nearest integer. ++ This means that if the digit left of the "point" (.) ++ is 1 we can add 0x80 to the mantissa since the ++ corner case 0x180 will round up to 0x200. If the ++ digit left of the "point" is 0 we will have to ++ add 0x7f since this will give 0xff and hence a ++ truncation/rounding downwards for the corner ++ case when the 9 lowest bits are 0x080 */ ++ mov \scratch, 0x7f /* Set rounding constant */ ++ /* Check if the mantissa is even or odd */ ++ bld \mant, 8 ++ subeq \scratch, -1 /* Rounding constant should be 0x80 */ ++ add \mant, \scratch ++ subcs \exp, -2 /* Adjust exponent if we overflowed */ ++.endm ++ ++/* Scale mantissa [mant_hi, mant_lo] with amount [shift_count]. ++ Uses scratch registers [scratch1] and [scratch2] */ ++.macro scale_df shift_count, mant_lo, mant_hi, scratch1, scratch2 ++ /* Scale [mant_hi, mant_lo] with shift_amount. ++ Must not forget the sticky bits we intend to shift out. */ ++ ++ rsub \scratch1,\shift_count,32/* get (32 - shift count) ++ (if shift count > 32 we get a ++ negative value, but that will ++ work as well in the code below.) */ ++ ++ cp.w \shift_count,32 /* handle shifts >= 32 separately */ ++ brhs 70f ++ ++ /* small (<32) shift amount, both words are part of the shift ++ first remember whether part that is lost contains any 1 bits ... */ ++ lsl \scratch2,\mant_lo,\scratch1 /*shift away bits that are part of ++ final mantissa. only part that goes ++ to scratch2 are bits that will be lost */ ++ ++ /* ... and now to the actual shift */ ++ lsl \scratch1,\mant_hi,\scratch1 /* get bits from msw destined for lsw*/ ++ lsr \mant_lo,\mant_lo,\shift_count /* shift down lsw of mantissa */ ++ lsr \mant_hi,\mant_hi,\shift_count /* shift down msw of mantissa */ ++ or \mant_lo,\scratch1 /* combine these bits with prepared lsw*/ ++ rjmp 71f ++ ++ /* large (>=32) shift amount, only lsw will have bits left after shift. ++ note that shift operations will use ((shift count) mod 32) so ++ we do not need to subtract 32 from shift count. */ ++70: ++ /* first remember whether part that is lost contains any 1 bits ... */ ++ lsl \scratch2,\mant_hi,\scratch1 /* save all lost bits from msw */ ++ or \scratch2,\mant_lo /* also save lost bits (all) from lsw ++ now scratch2<>0 if we lose any bits */ ++ ++ /* ... and now to the actual shift */ ++ mov \mant_lo,\mant_hi /* msw -> lsw (i.e. "shift 32 first")*/ ++ mov \mant_hi,0 /* clear msw */ ++ lsr \mant_lo,\mant_lo,\shift_count /* make rest of shift inside lsw*/ ++ ++71: ++ cp.w \scratch2,0 /* if any '1' bit in part we lost ...*/ ++ breq 70f ++ ++ sbr \mant_lo,0 /* ... we need to set sticky bit*/ ++70: ++.endm ++ ++/* Unpack exponent and mantissa from the double number ++ stored in [df_hi,df_lo]. The exponent is stored in [exp] ++ while the mantissa is stored in [df_hi,df_lo]. */ ++ ++.macro unpack_df exp, df_lo, df_hi ++ lsr \exp, \df_hi,21 /* Extract exponent */ ++ lsl \df_hi,10 /* Get mantissa */ ++ or \df_hi,\df_hi,\df_lo>>21 ++ lsl \df_lo,11 ++ ++ neg \exp /* Fix implicit bit */ ++ bst \df_hi,31 ++ subeq \exp,1 ++ neg \exp /* negate back exponent */ ++ .endm ++ ++/* Unpack exponent and mantissa from the single float number ++ stored in [sf]. The exponent is stored in [exp] ++ while the mantissa is stored in [sf]. */ ++.macro unpack_sf exp, sf ++ lsr \exp, \sf, 24 ++ brne 80f ++ /* Fix subnormal number */ ++ lsl \sf,7 ++ clz \exp,\sf ++ lsl \sf,\sf,\exp ++ rsub \exp,\exp,1 ++ rjmp 81f ++80: ++ lsl \sf,7 ++ sbr \sf, 31 /*Implicit bit*/ ++81: ++.endm ++ ++ ++ ++/* Pack a single float number stored in [mant] and [exp] ++ into a single float number in [sf] */ ++.macro pack_sf sf, exp, mant ++ bld \mant,31 /* implicit bit to z */ ++ subne \exp,1 /* if subnormal (implicit bit 0) ++ adjust exponent to storage format */ ++ ++ lsr \sf, \mant, 7 ++ bfins \sf, \exp, 24, 8 ++.endm ++ ++/* Pack exponent [exp] and mantissa [mant_hi, mant_lo] ++ into [df_hi, df_lo]. [df_hi] is shifted ++ one bit up so the sign bit can be shifted into it */ ++ ++.macro pack_df exp, mant_lo, mant_hi, df_lo, df_hi ++ bld \mant_hi,31 /* implicit bit to z */ ++ subne \exp,1 /* if subnormal (implicit bit 0) ++ adjust exponent to storage format */ ++ ++ lsr \mant_lo,11 /* shift back lsw */ ++ or \df_lo,\mant_lo,\mant_hi<<21 /* combine with low bits from msw */ ++ lsl \mant_hi,1 /* get rid of implicit bit */ ++ lsr \mant_hi,11 /* shift back msw except for one step*/ ++ or \df_hi,\mant_hi,\exp<<21 /* combine msw with exponent */ ++.endm ++ ++/* Normalize single float number stored in [mant] and [exp] ++ using scratch register [scratch] */ ++.macro normalize_sf exp, mant, scratch ++ /* Adjust exponent and mantissa */ ++ clz \scratch, \mant ++ sub \exp, \scratch ++ lsl \mant, \mant, \scratch ++.endm ++ ++/* Normalize the exponent and mantissa pair stored ++ in [mant_hi,mant_lo] and [exp]. Needs two scratch ++ registers [scratch1] and [scratch2]. */ ++.macro normalize_df exp, mant_lo, mant_hi, scratch1, scratch2 ++ clz \scratch1,\mant_hi /* Check if we have zeros in high bits */ ++ breq 80f /* No need for scaling if no zeros in high bits */ ++ brcs 81f /* Check for all zeros */ ++ ++ /* shift amount is smaller than 32, and involves both msw and lsw*/ ++ rsub \scratch2,\scratch1,32 /* shift mantissa */ ++ lsl \mant_hi,\mant_hi,\scratch1 ++ lsr \scratch2,\mant_lo,\scratch2 ++ or \mant_hi,\scratch2 ++ lsl \mant_lo,\mant_lo,\scratch1 ++ sub \exp,\scratch1 /* adjust exponent */ ++ rjmp 80f /* Finished */ ++81: ++ /* shift amount is greater than 32 */ ++ clz \scratch1,\mant_lo /* shift mantissa */ ++ movcs \scratch1, 0 ++ subcc \scratch1,-32 ++ mov \mant_hi,\mant_lo ++ lsl \mant_hi,\mant_hi,\scratch1 ++ mov \mant_lo,0 ++ sub \exp,\scratch1 /* adjust exponent */ ++80: ++.endm ++ ++ ++/* Fast but approximate multiply of two 64-bit numbers to give a 64 bit result. ++ The multiplication of [al]x[bl] is discarded. ++ Operands in [ah], [al], [bh], [bl]. ++ Scratch registers in [sh], [sl]. ++ Returns results in registers [rh], [rl].*/ ++.macro mul_approx_df ah, al, bh, bl, rh, rl, sh, sl ++ mulu.d \sl, \ah, \bl ++ macu.d \sl, \al, \bh ++ mulu.d \rl, \ah, \bh ++ add \rl, \sh ++ acr \rh ++.endm ++ ++ ++ ++ ++#ifdef L_avr32_f64_mul ++ .align 2 ++ .global __avr32_f64_mul ++ .type __avr32_f64_mul,@function ++ ++ ++__avr32_f64_mul: ++ or r12, r10, r11 << 1 ++ breq __avr32_f64_mul_op1_zero ++ ++ stm --sp, r5,r6,r7,lr ++ /* op1 in {r11,r10}*/ ++ /* op2 in {r9,r8}*/ ++ eor lr, r11, r9 /* MSB(lr) = Sign(op1) ^ Sign(op2) */ ++ ++ /* Unpack op1 */ ++ /* exp: r7 */ ++ /* sf: r11, r10 */ ++ lsr r7, r11, 20 /* Extract exponent */ ++ ++ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r11, r11, r10>>21 ++ lsl r10, 11 ++ sbr r11, 31 /* Insert implicit bit */ ++ ++ cbr r7, 11 /* Clear sign bit */ ++ /* Check if normalization is needed */ ++ breq __avr32_f64_mul_op1_subnormal /*If number is subnormal, normalize it */ ++ ++22: ++ /* Unpack op2 */ ++ /* exp: r6 */ ++ /* sf: r9, r8 */ ++ lsr r6, r9, 20 /* Extract exponent */ ++ ++ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>21 ++ lsl r8, 11 ++ sbr r9, 31 /* Insert implicit bit */ ++ ++ cbr r6, 11 /* Clear sign bit */ ++ /* Check if normalization is needed */ ++ breq __avr32_f64_mul_op2_subnormal /*If number is subnormal, normalize it */ ++23: ++ ++ /* Check if any operands are NaN or INF */ ++ cp r7, 0x7ff ++ breq __avr32_f64_mul_op_nan_or_inf /* Check op1 for NaN or Inf */ ++ cp r6, 0x7ff ++ breq __avr32_f64_mul_op_nan_or_inf /* Check op2 for NaN or Inf */ ++ ++ ++ /* Calculate new exponent in r12*/ ++ add r12, r7, r6 ++ sub r12, (1023-1) ++ ++ /* Do the multiplication using approximate calculation. ++ Place result in r11, r10. Use r7, r6 as scratch registers */ ++ mul_approx_df r11 /*ah*/, r10 /*al*/, r9 /*bh*/, r8 /*bl*/, r11 /*rh*/, r10 /*rl*/, r7 /*sh*/, r6 /*sl*/ ++ ++ /* Check if result is zero */ ++ breq __avr32_f64_mul_res_subnormal ++ ++ /* Adjust exponent and mantissa */ ++ /* [r12]:exp, [r11, r10]:mant, [r9,r8,r7]:scratch*/ ++ /* Mantissa may be of the format 0.xxxx or 1.xxxx. */ ++ /* In the first case, shift one pos to left.*/ ++ sub r9, r12, 1 ++ mov r8, r11 ++ lsl r7, r10, 1 ++ rol r8 ++ bld r11, 31 ++ movne r12, r9 ++ movne r11, r8 ++ movne r10, r7 ++ cp r12, 0 ++ breq __avr32_f64_mul_res_subnormal /*Result was subnormal. Flush-to-zero and return zero*/ ++ ++ /* Check for Inf. */ ++ cp.w r12, 0x7ff ++ brge __avr32_f64_mul_res_inf ++ ++ /* Result was not subnormal. Perform rounding. */ ++ /* Because of performance optimization, we have no sticky bit, */ ++ /* so round-to-even won't work as specified in the IEEE standard.*/ ++ /* [r12]:exp, [r11, r10]:mant */ ++ /* Mantissa is in 0.64 format. Round by adding 1<<(64-(52+2))=1<<10*/ ++ /* That is, 1 in the MSB of the part that will be discarded in final packing.*/ ++ mov r9, (1<<10) ++ add r10, r9 ++ acr r11 ++ /* Adjust exponent if we overflowed.*/ ++ subcs r12, -1 ++ ++ ++ /* Pack final result*/ ++ /* Input: [r12]:exp, [r11, r10]:mant */ ++ /* Result in [r11,r10] */ ++ /* Insert mantissa */ ++ cbr r11, 31 /*Clear implicit bit*/ ++ lsr r10, 11 ++ or r10, r10, r11<<21 ++ lsr r11, 11 ++ /* Insert exponent and sign bit*/ ++ or r11, r11, r12<<20 ++ bld lr, 31 ++ bst r11, 31 ++ ++ /* Return result in [r11,r10] */ ++ ldm sp++, r5, r6, r7,pc ++ ++ ++__avr32_f64_mul_op1_subnormal: ++ cbr r11, 31 /* Clear implicit bit. */ ++ normalize_df r7 /*exp*/, r10, r11 /*Mantissa*/, r5, r12 /*scratch*/ ++ rjmp 22b ++ ++__avr32_f64_mul_op2_subnormal: ++ cbr r9, 31 /* Clear implicit bit. */ ++ normalize_df r6 /*exp*/, r8, r9 /*Mantissa*/, r5, r12 /*scratch*/ ++ rjmp 23b ++ ++ ++__avr32_f64_mul_op_nan_or_inf: ++ /* Same code for OP1 and OP2*/ ++ /* Since we are here, at least one of the OPs were NaN or INF*/ ++ /* Shift out implicit bit of both operands' mantissa */ ++ lsl r11, 1 ++ lsl r9, 1 ++ /* Merge the regs in each operand to check for zero*/ ++ or r11, r10 /* op1 */ ++ or r9, r8 /* op2 */ ++ /* Check if op1 is NaN or INF */ ++ cp r7, 0x7ff ++ brne __avr32_f64_mul_op1_not_naninf ++ /* op1 was NaN or INF.*/ ++ cp r11, 0 ++ brne __avr32_f64_mul_res_nan /* op1 was NaN. Result will be NaN*/ ++ /*op1 was INF. check if op2 is NaN or INF*/ ++ cp r6, 0x7ff ++ brne __avr32_f64_mul_res_inf /*op1 was INF, op2 was neither NaN nor INF*/ ++ /* op1 is INF, op2 is either NaN or INF*/ ++ cp r9, 0 ++ breq __avr32_f64_mul_res_inf /*op2 was also INF*/ ++ rjmp __avr32_f64_mul_res_nan /*op2 was NaN*/ ++ ++__avr32_f64_mul_op1_not_naninf: ++ /* op1 was not NaN nor INF. Then op2 must be NaN or INF*/ ++ cp r9, 0 ++ breq __avr32_f64_mul_res_inf /*op2 was INF, return INF*/ ++ rjmp __avr32_f64_mul_res_nan /*else return NaN*/ ++ ++__avr32_f64_mul_res_subnormal:/* Multiply result was subnormal. Return zero. */ ++ mov r11, lr /*Get correct sign*/ ++ andh r11, 0x8000, COH ++ mov r10, 0 ++ ldm sp++, r5, r6, r7,pc ++ ++__avr32_f64_mul_res_nan: /* Return NaN. */ ++ mov r11, -1 ++ mov r10, -1 ++ ldm sp++, r5, r6, r7,pc ++ ++__avr32_f64_mul_res_inf: /* Return INF. */ ++ mov r11, 0 ++ orh r11, 0x7ff0 ++ bld lr, 31 ++ bst r11, 31 ++ mov r10, 0 ++ ldm sp++, r5, r6, r7,pc ++ ++__avr32_f64_mul_op1_zero: ++ /* Get sign */ ++ eor r11, r11, r9 ++ andh r11, 0x8000, COH ++ /* Check if op2 is Inf or NaN. */ ++ bfextu r12, r9, 20, 11 ++ cp.w r12, 0x7ff ++ retne r12 /* Return 0.0 */ ++ /* Return NaN */ ++ mov r10, -1 ++ mov r11, -1 ++ ret r12 ++ ++ ++ ++#endif ++ ++ ++#if defined(L_avr32_f64_addsub) || defined(L_avr32_f64_addsub_fast) ++ .align 2 ++ ++__avr32_f64_sub_from_add: ++ /* Switch sign on op2 */ ++ eorh r9, 0x8000 ++ ++#if defined(L_avr32_f64_addsub_fast) ++ .global __avr32_f64_sub_fast ++ .type __avr32_f64_sub_fast,@function ++__avr32_f64_sub_fast: ++#else ++ .global __avr32_f64_sub ++ .type __avr32_f64_sub,@function ++__avr32_f64_sub: ++#endif ++ ++ /* op1 in {r11,r10}*/ ++ /* op2 in {r9,r8}*/ ++ ++#if defined(L_avr32_f64_addsub_fast) ++ /* If op2 is zero just return op1 */ ++ or r12, r8, r9 << 1 ++ reteq r12 ++#endif ++ ++ /* Check signs */ ++ eor r12, r11, r9 ++ /* Different signs, use addition. */ ++ brmi __avr32_f64_add_from_sub ++ ++ stm --sp, r5, r6, r7, lr ++ ++ /* Get sign of op1 into r12 */ ++ mov r12, r11 ++ andh r12, 0x8000, COH ++ ++ /* Remove sign from operands */ ++ cbr r11, 31 ++ cbr r9, 31 ++ ++ /* Put the number with the largest exponent in [r11, r10] ++ and the number with the smallest exponent in [r9, r8] */ ++ cp r11, r9 ++ brhs 1f /* Skip swap if operands already correctly ordered*/ ++ /* Operands were not correctly ordered, swap them*/ ++ mov r7, r11 ++ mov r11, r9 ++ mov r9, r7 ++ mov r7, r10 ++ mov r10, r8 ++ mov r8, r7 ++ eorh r12, 0x8000 /* Invert sign in r12*/ ++1: ++ /* Unpack largest operand - opH */ ++ /* exp: r7 */ ++ /* sf: r11, r10 */ ++ lsr r7, r11, 20 /* Extract exponent */ ++ cbr r7, 11 /* Clear sign bit */ ++ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r11, r11, r10>>21 ++ lsl r10, 11 ++ sbr r11, 31 /* Insert implicit bit */ ++ ++ ++ /* Unpack smallest operand - opL */ ++ /* exp: r6 */ ++ /* sf: r9, r8 */ ++ lsr r6, r9, 20 /* Extract exponent */ ++ cbr r6, 11 /* Clear sign bit */ ++ breq __avr32_f64_sub_opL_subnormal /* If either zero or subnormal */ ++ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>21 ++ lsl r8, 11 ++ sbr r9, 31 /* Insert implicit bit */ ++ ++ ++__avr32_f64_sub_opL_subnormal_done: ++ /* opH is NaN or Inf. */ ++ cp.w r7, 0x7ff ++ breq __avr32_f64_sub_opH_nan_or_inf ++ ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r6, r7 ++ breq __avr32_f64_sub_shift_done /* No need to shift, exponents are equal*/ ++ ++ /* Scale mantissa [r9, r8] with amount [r6]. ++ Uses scratch registers [r5] and [lr]. ++ In IEEE mode:Must not forget the sticky bits we intend to shift out. */ ++ ++ rsub r5,r6,32 /* get (32 - shift count) ++ (if shift count > 32 we get a ++ negative value, but that will ++ work as well in the code below.) */ ++ ++ cp.w r6,32 /* handle shifts >= 32 separately */ ++ brhs __avr32_f64_sub_longshift ++ ++ /* small (<32) shift amount, both words are part of the shift ++ first remember whether part that is lost contains any 1 bits ... */ ++ lsl lr,r8,r5 /* shift away bits that are part of ++ final mantissa. only part that goes ++ to lr are bits that will be lost */ ++ ++ /* ... and now to the actual shift */ ++ lsl r5,r9,r5 /* get bits from msw destined for lsw*/ ++ lsr r8,r8,r6 /* shift down lsw of mantissa */ ++ lsr r9,r9,r6 /* shift down msw of mantissa */ ++ or r8,r5 /* combine these bits with prepared lsw*/ ++#if defined(L_avr32_f64_addsub) ++ cp.w lr,0 /* if any '1' bit in part we lost ...*/ ++ srne lr ++ or r8, lr /* ... we need to set sticky bit*/ ++#endif ++ ++__avr32_f64_sub_shift_done: ++ /* Now subtract the mantissas. */ ++ sub r10, r8 ++ sbc r11, r11, r9 ++ ++ /* Normalize the exponent and mantissa pair stored in ++ [r11,r10] and exponent in [r7]. Needs two scratch registers [r6] and [lr]. */ ++ clz r6,r11 /* Check if we have zeros in high bits */ ++ breq __avr32_f64_sub_longnormalize_done /* No need for scaling if no zeros in high bits */ ++ brcs __avr32_f64_sub_longnormalize ++ ++ /* shift amount is smaller than 32, and involves both msw and lsw*/ ++ rsub lr,r6,32 /* shift mantissa */ ++ lsl r11,r11,r6 ++ lsr lr,r10,lr ++ or r11,lr ++ lsl r10,r10,r6 ++ sub r7,r6 /* adjust exponent */ ++ brle __avr32_f64_sub_subnormal_result ++ ++__avr32_f64_sub_longnormalize_done: ++ ++#if defined(L_avr32_f64_addsub) ++ /* Insert the bits we will remove from the mantissa r9[31:21] */ ++ lsl r9, r10, (32 - 11) ++#else ++ /* Keep the last bit shifted out. */ ++ bfextu r9, r10, 10, 1 ++#endif ++ ++ /* Pack final result*/ ++ /* Input: [r7]:exp, [r11, r10]:mant, [r12]:sign in MSB */ ++ /* Result in [r11,r10] */ ++ /* Insert mantissa */ ++ cbr r11, 31 /*Clear implicit bit*/ ++ lsr r10, 11 ++ or r10, r10, r11<<21 ++ lsr r11, 11 ++ /* Insert exponent and sign bit*/ ++ or r11, r11, r7<<20 ++ bld r12, 31 ++ bst r11, 31 ++ ++ /* Round */ ++__avr32_f64_sub_round: ++#if defined(L_avr32_f64_addsub) ++ mov r7, 0 ++ sbr r7, 31 ++ bld r10, 0 ++ subne r7, -1 ++ ++ cp.w r9, r7 ++ srhs r9 ++#endif ++ add r10, r9 ++ acr r11 ++ ++ /* Return result in [r11,r10] */ ++ ldm sp++, r5, r6, r7,pc ++ ++ ++ ++__avr32_f64_sub_opL_subnormal: ++ /* Extract the of mantissa */ ++ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>21 ++ lsl r8, 11 ++ ++ /* Set exponent to 1 if we do not have a zero. */ ++ or lr, r9, r8 ++ movne r6,1 ++ ++ /* Check if opH is also subnormal. If so, clear implicit bit in r11*/ ++ rsub lr, r7, 0 ++ moveq r7,1 ++ bst r11, 31 ++ ++ /* Check if op1 is zero, if so set exponent to 0. */ ++ or lr, r11, r10 ++ moveq r7,0 ++ ++ rjmp __avr32_f64_sub_opL_subnormal_done ++ ++__avr32_f64_sub_opH_nan_or_inf: ++ /* Check if opH is NaN, if so return NaN */ ++ cbr r11, 31 ++ or lr, r11, r10 ++ brne __avr32_f64_sub_return_nan ++ ++ /* opH is Inf. */ ++ /* Check if opL is Inf. or NaN */ ++ cp.w r6, 0x7ff ++ breq __avr32_f64_sub_opL_nan_or_inf ++ ldm sp++, r5, r6, r7, pc/* opL not Inf or NaN, return opH */ ++__avr32_f64_sub_opL_nan_or_inf: ++ cbr r9, 31 ++ or lr, r9, r8 ++ brne __avr32_f64_sub_return_nan ++ mov r10, 0 /* Generate Inf in r11, r10 */ ++ mov r11, 0 ++ orh r11, 0x7ff0 ++ ldm sp++, r5, r6, r7, pc/* opL Inf, return Inf */ ++__avr32_f64_sub_return_nan: ++ mov r10, -1 /* Generate NaN in r11, r10 */ ++ mov r11, -1 ++ ldm sp++, r5, r6, r7, pc/* opL Inf or NaN, return NaN */ ++ ++ ++__avr32_f64_sub_subnormal_result: ++ /* Just flush subnormals to zero. */ ++ mov r10, 0 ++ mov r11, 0 ++ ldm sp++, r5, r6, r7, pc ++ ++ ++__avr32_f64_sub_longshift: ++ /* large (>=32) shift amount, only lsw will have bits left after shift. ++ note that shift operations will use ((shift count=r6) mod 32) so ++ we do not need to subtract 32 from shift count. */ ++ /* Saturate the shift amount to 63. If the amount ++ is any larger op2 is insignificant. */ ++ satu r6 >> 0, 6 ++#if defined(L_avr32_f64_addsub) ++ /* first remember whether part that is lost contains any 1 bits ... */ ++ lsl lr,r9,r5 /* save all lost bits from msw */ ++ or lr,r8 /* also save lost bits (all) from lsw ++ now lr != 0 if we lose any bits */ ++#endif ++ /* ... and now to the actual shift */ ++ mov r8,r9 /* msw -> lsw (i.e. "shift 32 first")*/ ++ mov r9,0 /* clear msw */ ++ lsr r8,r8,r6 /* make rest of shift inside lsw*/ ++#if defined(L_avr32_f64_addsub) ++ cp.w lr,0 /* if any '1' bit in part we lost ...*/ ++ sreq lr ++ or r8, lr /* ... we need to set sticky bit*/ ++#endif ++ rjmp __avr32_f64_sub_shift_done ++ ++__avr32_f64_sub_longnormalize: ++ /* shift amount is greater than 32 */ ++ clz r6,r10 /* shift mantissa */ ++ /* If the resulting mantissa is zero the result is ++ zero so force exponent to zero. */ ++ movcs r7, 0 ++ movcs r6, 0 ++ subcc r6,-32 ++ mov r11,r10 ++ lsl r11,r11,r6 ++ mov r10,0 ++ sub r7,r6 /* adjust exponent */ ++ brle __avr32_f64_sub_subnormal_result ++ rjmp __avr32_f64_sub_longnormalize_done ++ ++ ++ ++ .align 2 ++__avr32_f64_add_from_sub: ++ /* Switch sign on op2 */ ++ eorh r9, 0x8000 ++ ++#if defined(L_avr32_f64_addsub_fast) ++ .global __avr32_f64_add_fast ++ .type __avr32_f64_add_fast,@function ++__avr32_f64_add_fast: ++#else ++ .global __avr32_f64_add ++ .type __avr32_f64_add,@function ++__avr32_f64_add: ++#endif ++ ++ /* op1 in {r11,r10}*/ ++ /* op2 in {r9,r8}*/ ++ ++#if defined(L_avr32_f64_addsub_fast) ++ /* If op2 is zero just return op1 */ ++ or r12, r8, r9 << 1 ++ reteq r12 ++#endif ++ ++ /* Check signs */ ++ eor r12, r11, r9 ++ /* Different signs, use subtraction. */ ++ brmi __avr32_f64_sub_from_add ++ ++ stm --sp, r5, r6, r7, lr ++ ++ /* Get sign of op1 into r12 */ ++ mov r12, r11 ++ andh r12, 0x8000, COH ++ ++ /* Remove sign from operands */ ++ cbr r11, 31 ++ cbr r9, 31 ++ ++ /* Put the number with the largest exponent in [r11, r10] ++ and the number with the smallest exponent in [r9, r8] */ ++ cp r11, r9 ++ brhs 1f /* Skip swap if operands already correctly ordered*/ ++ /* Operands were not correctly ordered, swap them*/ ++ mov r7, r11 ++ mov r11, r9 ++ mov r9, r7 ++ mov r7, r10 ++ mov r10, r8 ++ mov r8, r7 ++1: ++ /* Unpack largest operand - opH */ ++ /* exp: r7 */ ++ /* sf: r11, r10 */ ++ lsr r7, r11, 20 /* Extract exponent */ ++ cbr r7, 11 /* Clear sign bit */ ++ lsl r11, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r11, r11, r10>>21 ++ lsl r10, 11 ++ sbr r11, 31 /* Insert implicit bit */ ++ ++ ++ /* Unpack smallest operand - opL */ ++ /* exp: r6 */ ++ /* sf: r9, r8 */ ++ lsr r6, r9, 20 /* Extract exponent */ ++ cbr r6, 11 /* Clear sign bit */ ++ breq __avr32_f64_add_opL_subnormal /* If either zero or subnormal */ ++ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>21 ++ lsl r8, 11 ++ sbr r9, 31 /* Insert implicit bit */ ++ ++ ++__avr32_f64_add_opL_subnormal_done: ++ /* opH is NaN or Inf. */ ++ cp.w r7, 0x7ff ++ breq __avr32_f64_add_opH_nan_or_inf ++ ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r6, r7 ++ breq __avr32_f64_add_shift_done /* No need to shift, exponents are equal*/ ++ ++ /* Scale mantissa [r9, r8] with amount [r6]. ++ Uses scratch registers [r5] and [lr]. ++ In IEEE mode:Must not forget the sticky bits we intend to shift out. */ ++ rsub r5,r6,32 /* get (32 - shift count) ++ (if shift count > 32 we get a ++ negative value, but that will ++ work as well in the code below.) */ ++ ++ cp.w r6,32 /* handle shifts >= 32 separately */ ++ brhs __avr32_f64_add_longshift ++ ++ /* small (<32) shift amount, both words are part of the shift ++ first remember whether part that is lost contains any 1 bits ... */ ++ lsl lr,r8,r5 /* shift away bits that are part of ++ final mantissa. only part that goes ++ to lr are bits that will be lost */ ++ ++ /* ... and now to the actual shift */ ++ lsl r5,r9,r5 /* get bits from msw destined for lsw*/ ++ lsr r8,r8,r6 /* shift down lsw of mantissa */ ++ lsr r9,r9,r6 /* shift down msw of mantissa */ ++ or r8,r5 /* combine these bits with prepared lsw*/ ++#if defined(L_avr32_f64_addsub) ++ cp.w lr,0 /* if any '1' bit in part we lost ...*/ ++ srne lr ++ or r8, lr /* ... we need to set sticky bit*/ ++#endif ++ ++__avr32_f64_add_shift_done: ++ /* Now add the mantissas. */ ++ add r10, r8 ++ adc r11, r11, r9 ++ ++ /* Check if we overflowed. */ ++ brcs __avr32_f64_add_res_of ++ ++__avr32_f64_add_res_of_done: ++ ++#if defined(L_avr32_f64_addsub) ++ /* Insert the bits we will remove from the mantissa r9[31:21] */ ++ lsl r9, r10, (32 - 11) ++#else ++ /* Keep the last bit shifted out. */ ++ bfextu r9, r10, 10, 1 ++#endif ++ ++ /* Pack final result*/ ++ /* Input: [r7]:exp, [r11, r10]:mant, [r12]:sign in MSB */ ++ /* Result in [r11,r10] */ ++ /* Insert mantissa */ ++ cbr r11, 31 /*Clear implicit bit*/ ++ lsr r10, 11 ++ or r10, r10, r11<<21 ++ lsr r11, 11 ++ /* Insert exponent and sign bit*/ ++ or r11, r11, r7<<20 ++ bld r12, 31 ++ bst r11, 31 ++ ++ /* Round */ ++__avr32_f64_add_round: ++#if defined(L_avr32_f64_addsub) ++ mov r7, 0 ++ sbr r7, 31 ++ bld r10, 0 ++ subne r7, -1 ++ ++ cp.w r9, r7 ++ srhs r9 ++#endif ++ add r10, r9 ++ acr r11 ++ ++ /* Return result in [r11,r10] */ ++ ldm sp++, r5, r6, r7,pc ++ ++ ++ ++__avr32_f64_add_opL_subnormal: ++ /* Extract the of mantissa */ ++ lsl r9, 11 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>21 ++ lsl r8, 11 ++ ++ /* Set exponent to 1 if we do not have a zero. */ ++ or lr, r9, r8 ++ movne r6,1 ++ ++ /* Check if opH is also subnormal. If so, clear implicit bit in r11*/ ++ rsub lr, r7, 0 ++ moveq r7,1 ++ bst r11, 31 ++ ++ /* Check if op1 is zero, if so set exponent to 0. */ ++ or lr, r11, r10 ++ moveq r7,0 ++ ++ rjmp __avr32_f64_add_opL_subnormal_done ++ ++__avr32_f64_add_opH_nan_or_inf: ++ /* Check if opH is NaN, if so return NaN */ ++ cbr r11, 31 ++ or lr, r11, r10 ++ brne __avr32_f64_add_return_nan ++ ++ /* opH is Inf. */ ++ /* Check if opL is Inf. or NaN */ ++ cp.w r6, 0x7ff ++ breq __avr32_f64_add_opL_nan_or_inf ++ ldm sp++, r5, r6, r7, pc/* opL not Inf or NaN, return opH */ ++__avr32_f64_add_opL_nan_or_inf: ++ cbr r9, 31 ++ or lr, r9, r8 ++ brne __avr32_f64_add_return_nan ++ mov r10, 0 /* Generate Inf in r11, r10 */ ++ mov r11, 0 ++ orh r11, 0x7ff0 ++ ldm sp++, r5, r6, r7, pc/* opL Inf, return Inf */ ++__avr32_f64_add_return_nan: ++ mov r10, -1 /* Generate NaN in r11, r10 */ ++ mov r11, -1 ++ ldm sp++, r5, r6, r7, pc/* opL Inf or NaN, return NaN */ ++ ++ ++__avr32_f64_add_longshift: ++ /* large (>=32) shift amount, only lsw will have bits left after shift. ++ note that shift operations will use ((shift count=r6) mod 32) so ++ we do not need to subtract 32 from shift count. */ ++ /* Saturate the shift amount to 63. If the amount ++ is any larger op2 is insignificant. */ ++ satu r6 >> 0, 6 ++#if defined(L_avr32_f64_addsub) ++ /* first remember whether part that is lost contains any 1 bits ... */ ++ lsl lr,r9,r5 /* save all lost bits from msw */ ++ or lr,r8 /* also save lost bits (all) from lsw ++ now lr != 0 if we lose any bits */ ++#endif ++ /* ... and now to the actual shift */ ++ mov r8,r9 /* msw -> lsw (i.e. "shift 32 first")*/ ++ mov r9,0 /* clear msw */ ++ lsr r8,r8,r6 /* make rest of shift inside lsw*/ ++#if defined(L_avr32_f64_addsub) ++ cp.w lr,0 /* if any '1' bit in part we lost ...*/ ++ sreq lr ++ or r8, lr /* ... we need to set sticky bit*/ ++#endif ++ rjmp __avr32_f64_add_shift_done ++ ++__avr32_f64_add_res_of: ++ /* We overflowed. Increase exponent and shift mantissa.*/ ++ /* [r7]:exp, [r11, r10]:mant */ ++ ror r11 ++ lsr r10, 1 ++ sub r7, -1 ++ ++ /* Clear mantissa to set result to Inf if the exponent is 255. */ ++ cp.w r7, 0x7ff ++ moveq r10, 0 ++ moveq r11, 0 ++ rjmp __avr32_f64_add_res_of_done ++ ++ ++#endif ++ ++#ifdef L_avr32_f64_to_u32 ++ /* This goes into L_fixdfsi */ ++#endif ++ ++ ++#ifdef L_avr32_f64_to_s32 ++ .global __avr32_f64_to_u32 ++ .type __avr32_f64_to_u32,@function ++__avr32_f64_to_u32: ++ cp.w r11, 0 ++ retmi 0 /* Negative returns 0 */ ++ ++ /* Fallthrough to df to signed si conversion */ ++ .global __avr32_f64_to_s32 ++ .type __avr32_f64_to_s32,@function ++__avr32_f64_to_s32: ++ lsl r12,r11,1 ++ lsr r12,21 /* extract exponent*/ ++ sub r12,1023 /* convert to unbiased exponent.*/ ++ retlo 0 /* too small exponent implies zero. */ ++ ++1: ++ rsub r12,r12,31 /* shift count = 31 - exponent */ ++ mov r9,r11 /* save sign for later...*/ ++ lsl r11,11 /* remove exponent and sign*/ ++ sbr r11,31 /* add implicit bit*/ ++ or r11,r11,r10>>21 /* get rest of bits from lsw of double */ ++ lsr r11,r11,r12 /* shift down mantissa to final place */ ++ lsl r9,1 /* sign -> carry */ ++ retcc r11 /* if positive, we are done */ ++ neg r11 /* if negative float, negate result */ ++ ret r11 ++ ++#endif /* L_fixdfsi*/ ++ ++#ifdef L_avr32_f64_to_u64 ++ /* Actual function is in L_fixdfdi */ ++#endif ++ ++#ifdef L_avr32_f64_to_s64 ++ .global __avr32_f64_to_u64 ++ .type __avr32_f64_to_u64,@function ++__avr32_f64_to_u64: ++ cp.w r11,0 ++ /* Negative numbers return zero */ ++ movmi r10, 0 ++ movmi r11, 0 ++ retmi r11 ++ ++ ++ ++ /* Fallthrough */ ++ .global __avr32_f64_to_s64 ++ .type __avr32_f64_to_s64,@function ++__avr32_f64_to_s64: ++ lsl r9,r11,1 ++ lsr r9,21 /* get exponent*/ ++ sub r9,1023 /* convert to correct range*/ ++ /* Return zero if exponent to small */ ++ movlo r10, 0 ++ movlo r11, 0 ++ retlo r11 ++ ++ mov r8,r11 /* save sign for later...*/ ++1: ++ lsl r11,11 /* remove exponent */ ++ sbr r11,31 /* add implicit bit*/ ++ or r11,r11,r10>>21 /* get rest of bits from lsw of double*/ ++ lsl r10,11 /* align lsw correctly as well */ ++ rsub r9,r9,63 /* shift count = 63 - exponent */ ++ breq 1f ++ ++ cp.w r9,32 /* is shift count more than one reg? */ ++ brhs 0f ++ ++ mov r12,r11 /* save msw */ ++ lsr r10,r10,r9 /* small shift count, shift down lsw */ ++ lsr r11,r11,r9 /* small shift count, shift down msw */ ++ rsub r9,r9,32 /* get 32-size of shifted out tail */ ++ lsl r12,r12,r9 /* align part to move from msw to lsw */ ++ or r10,r12 /* combine to get new lsw */ ++ rjmp 1f ++ ++0: ++ lsr r10,r11,r9 /* large shift count,only lsw get bits ++ note that shift count is modulo 32*/ ++ mov r11,0 /* msw will be 0 */ ++ ++1: ++ lsl r8,1 /* sign -> carry */ ++ retcc r11 /* if positive, we are done */ ++ ++ neg r11 /* if negative float, negate result */ ++ neg r10 ++ scr r11 ++ ret r11 ++ ++#endif ++ ++#ifdef L_avr32_u32_to_f64 ++ /* Code located in L_floatsidf */ ++#endif ++ ++#ifdef L_avr32_s32_to_f64 ++ .global __avr32_u32_to_f64 ++ .type __avr32_u32_to_f64,@function ++__avr32_u32_to_f64: ++ sub r11, r12, 0 /* Move to r11 and force Z flag to be updated */ ++ mov r12, 0 /* always positive */ ++ rjmp 0f /* Jump to common code for floatsidf */ ++ ++ .global __avr32_s32_to_f64 ++ .type __avr32_s32_to_f64,@function ++__avr32_s32_to_f64: ++ mov r11, r12 /* Keep original value in r12 for sign */ ++ abs r11 /* Absolute value if r12 */ ++0: ++ mov r10,0 /* let remaining bits be zero */ ++ reteq r11 /* zero long will return zero float */ ++ ++ pushm lr ++ mov r9,31+1023 /* set exponent */ ++ ++ normalize_df r9 /*exp*/, r10, r11 /* mantissa */, r8, lr /* scratch */ ++ ++ /* Check if a subnormal result was created */ ++ cp.w r9, 0 ++ brgt 0f ++ ++ adjust_subnormal_df r9 /* exp */, r10, r11 /* Mantissa */, r12 /*sign*/, r8, lr /* scratch */ ++ popm pc ++0: ++ ++ /* Round result */ ++ round_df r9 /*exp*/, r10, r11 /* Mantissa */, r8 /*scratch*/ ++ cp.w r9,0x7ff ++ brlt 0f ++ /*Return infinity */ ++ mov r10, 0 ++ mov r11, 0 ++ orh r11, 0xffe0 ++ rjmp __floatsidf_return_op1 ++ ++0: ++ ++ /* Pack */ ++ pack_df r9 /*exp*/, r10, r11 /* mantissa */, r10, r11 /* Output df number*/ ++__floatsidf_return_op1: ++ lsl r12,1 /* shift in sign bit */ ++ ror r11 ++ ++ popm pc ++#endif ++ ++ ++#ifdef L_avr32_f32_cmp_eq ++ .global __avr32_f32_cmp_eq ++ .type __avr32_f32_cmp_eq,@function ++__avr32_f32_cmp_eq: ++ cp.w r12, r11 ++ breq 0f ++ /* If not equal check for +/-0 */ ++ /* Or together the two values and shift out the sign bit. ++ If the result is zero, then the two values are both zero. */ ++ or r12, r11 ++ lsl r12, 1 ++ sreq r12 ++ ret r12 ++0: ++ /* Numbers were equal. Check for NaN or Inf */ ++ mov r11, 0 ++ orh r11, 0xff00 ++ lsl r12, 1 ++ cp.w r12, r11 ++ srls r12 /* 0 if NaN, 1 otherwise */ ++ ret r12 ++#endif ++ ++#if defined(L_avr32_f32_cmp_ge) || defined(L_avr32_f32_cmp_lt) ++#ifdef L_avr32_f32_cmp_ge ++ .global __avr32_f32_cmp_ge ++ .type __avr32_f32_cmp_ge,@function ++__avr32_f32_cmp_ge: ++#endif ++#ifdef L_avr32_f32_cmp_lt ++ .global __avr32_f32_cmp_lt ++ .type __avr32_f32_cmp_lt,@function ++__avr32_f32_cmp_lt: ++#endif ++ lsl r10, r12, 1 /* Remove sign bits */ ++ lsl r9, r11, 1 ++ mov r8, 0 ++ orh r8, 0xff00 ++ cp.w r10, r8 ++ rethi 0 /* Op0 is NaN */ ++ cp.w r9, r8 ++ rethi 0 /* Op1 is Nan */ ++ ++ eor r8, r11, r12 ++ bld r12, 31 ++#ifdef L_avr32_f32_cmp_ge ++ srcc r8 /* Set result to true if op0 is positive*/ ++#endif ++#ifdef L_avr32_f32_cmp_lt ++ srcs r8 /* Set result to true if op0 is negative*/ ++#endif ++ retmi r8 /* Return if signs are different */ ++ brcs 0f /* Both signs negative? */ ++ ++ /* Both signs positive */ ++ cp.w r12, r11 ++#ifdef L_avr32_f32_cmp_ge ++ srhs r12 ++#endif ++#ifdef L_avr32_f32_cmp_lt ++ srlo r12 ++#endif ++ retal r12 ++0: ++ /* Both signs negative */ ++ cp.w r11, r12 ++#ifdef L_avr32_f32_cmp_ge ++ srhs r12 ++#endif ++#ifdef L_avr32_f32_cmp_lt ++ srlo r12 ++#endif ++ retal r12 ++#endif ++ ++ ++#ifdef L_avr32_f64_cmp_eq ++ .global __avr32_f64_cmp_eq ++ .type __avr32_f64_cmp_eq,@function ++__avr32_f64_cmp_eq: ++ cp.w r10,r8 ++ cpc r11,r9 ++ breq 0f ++ ++ /* Args were not equal*/ ++ /* Both args could be zero with different sign bits */ ++ lsl r11,1 /* get rid of sign bits */ ++ lsl r9,1 ++ or r11,r10 /* Check if all bits are zero */ ++ or r11,r9 ++ or r11,r8 ++ sreq r12 /* If all zeros the arguments are equal ++ so return 1 else return 0 */ ++ ret r12 ++0: ++ /* check for NaN */ ++ lsl r11,1 ++ mov r12, 0 ++ orh r12, 0xffe0 ++ cp.w r10,0 ++ cpc r11,r12 /* check if nan or inf */ ++ srls r12 /* If Arg is NaN return 0 else 1*/ ++ ret r12 /* Return */ ++ ++#endif ++ ++ ++#if defined(L_avr32_f64_cmp_ge) || defined(L_avr32_f64_cmp_lt) ++ ++#ifdef L_avr32_f64_cmp_ge ++ .global __avr32_f64_cmp_ge ++ .type __avr32_f64_cmp_ge,@function ++__avr32_f64_cmp_ge: ++#endif ++#ifdef L_avr32_f64_cmp_lt ++ .global __avr32_f64_cmp_lt ++ .type __avr32_f64_cmp_lt,@function ++__avr32_f64_cmp_lt: ++#endif ++ ++ /* compare magnitude of op1 and op2 */ ++ pushm lr ++ ++ lsl r11,1 /* Remove sign bit of op1 */ ++ srcs lr /* Sign op1 to lsb of lr*/ ++ lsl r9,1 /* Remove sign bit of op2 */ ++ rol lr /* Sign op2 to lsb of lr, sign bit op1 bit 1 of lr*/ ++ ++ /* Check for Nan */ ++ mov r12, 0 ++ orh r12, 0xffe0 ++ cp.w r10,0 ++ cpc r11,r12 ++ movhi r12, 0 /* Return false for NaN */ ++ brhi 0f /* We have NaN */ ++ cp.w r8,0 ++ cpc r9,r12 ++ movhi r12, 0 /* Return false for NaN */ ++ brhi 0f /* We have NaN */ ++ ++ cp.w lr,3 /* both operands negative ?*/ ++ breq 1f ++ ++ cp.w lr,1 /* both operands positive? */ ++ brlo 2f ++ ++ /* Different signs. If sign of op1 is negative the difference ++ between op1 and op2 will always be negative, and if op1 is ++ positive the difference will always be positive */ ++#ifdef L_avr32_f64_cmp_ge ++ sreq r12 ++#endif ++#ifdef L_avr32_f64_cmp_lt ++ srne r12 ++#endif ++ popm pc ++ ++ ++2: ++ /* Both operands positive. Just compute the difference */ ++ cp.w r10,r8 ++ cpc r11,r9 ++#ifdef L_avr32_f64_cmp_ge ++ srhs r12 ++#endif ++#ifdef L_avr32_f64_cmp_lt ++ srlo r12 ++#endif ++ popm pc ++ ++1: ++ /* Both operands negative. Compute the difference with operands switched */ ++ cp r8,r10 ++ cpc r9,r11 ++#ifdef L_avr32_f64_cmp_ge ++ srhs r12 ++#endif ++#ifdef L_avr32_f64_cmp_lt ++ srlo r12 ++#endif ++0: ++ popm pc ++#endif ++ ++ ++ ++#ifdef L_avr32_f64_div ++ .align 2 ++ .global __avr32_f64_div ++ .type __avr32_f64_div,@function ++ ++__avr32_f64_div: ++ stm --sp, r0, r1, r2, r3, r4, r5, r6, r7,lr ++ /* op1 in {r11,r10}*/ ++ /* op2 in {r9,r8}*/ ++ eor lr, r11, r9 /* MSB(lr) = Sign(op1) ^ Sign(op2) */ ++ ++ ++ /* Unpack op1 to 2.62 format*/ ++ /* exp: r7 */ ++ /* sf: r11, r10 */ ++ lsr r7, r11, 20 /* Extract exponent */ ++ ++ lsl r11, 9 /* Extract mantissa, leave room for implicit bit */ ++ or r11, r11, r10>>23 ++ lsl r10, 9 ++ sbr r11, 29 /* Insert implicit bit */ ++ andh r11, 0x3fff /*Mask last part of exponent since we use 2.62 format*/ ++ ++ cbr r7, 11 /* Clear sign bit */ ++ /* Check if normalization is needed */ ++ breq 11f /*If number is subnormal, normalize it */ ++22: ++ cp r7, 0x7ff ++ brhs 2f /* Check op1 for NaN or Inf */ ++ ++ /* Unpack op2 to 2.62 format*/ ++ /* exp: r6 */ ++ /* sf: r9, r8 */ ++ lsr r6, r9, 20 /* Extract exponent */ ++ ++ lsl r9, 9 /* Extract mantissa, leave room for implicit bit */ ++ or r9, r9, r8>>23 ++ lsl r8, 9 ++ sbr r9, 29 /* Insert implicit bit */ ++ andh r9, 0x3fff /*Mask last part of exponent since we use 2.62 format*/ ++ ++ cbr r6, 11 /* Clear sign bit */ ++ /* Check if normalization is needed */ ++ breq 13f /*If number is subnormal, normalize it */ ++23: ++ cp r6, 0x7ff ++ brhs 3f /* Check op2 for NaN or Inf */ ++ ++ /* Calculate new exponent */ ++ sub r7, r6 ++ sub r7,-1023 ++ ++ /* Divide */ ++ /* Approximating 1/d with the following recurrence: */ ++ /* R[j+1] = R[j]*(2-R[j]*d) */ ++ /* Using 2.62 format */ ++ /* TWO: r12 */ ++ /* d = op2 = divisor (2.62 format): r9,r8 */ ++ /* Multiply result : r5, r4 */ ++ /* Initial guess : r3, r2 */ ++ /* New approximations : r3, r2 */ ++ /* op1 = Dividend (2.62 format) : r11, r10 */ ++ ++ mov r12, 1 /* Load TWO */ ++ brev r12 ++ ++ /* Load initial guess, using look-up table */ ++ /* Initial guess is of format 01.XY, where XY is constructed as follows: */ ++ /* Let d be of following format: 00.1xy....., then XY=~xy */ ++ /* For d=00.100 = 0,5 -> initial guess=01.11 = 1,75 */ ++ /* For d=00.101 = 0,625 -> initial guess=01.11 = 1,5 */ ++ /* For d=00.110 = 0,75 -> initial guess=01.11 = 1,25 */ ++ /* For d=00.111 = 0,875 -> initial guess=01.11 = 1,0 */ ++ /* r2 is also part of the reg pair forming initial guess, but it*/ ++ /* is kept uninitialized to save one cycle since it has so low significance*/ ++ ++ lsr r3, r12, 1 ++ bfextu r4, r9, 27, 2 ++ com r4 ++ bfins r3, r4, 28, 2 ++ ++ /* First approximation */ ++ /* Approximating to 32 bits */ ++ /* r5 = R[j]*d */ ++ mulu.d r4, r3, r9 ++ /* r5 = 2-R[j]*d */ ++ sub r5, r12, r5<<2 ++ /* r3 = R[j]*(2-R[j]*d) */ ++ mulu.d r4, r3, r5 ++ lsl r3, r5, 2 ++ ++ /* Second approximation */ ++ /* Approximating to 32 bits */ ++ /* r5 = R[j]*d */ ++ mulu.d r4, r3, r9 ++ /* r5 = 2-R[j]*d */ ++ sub r5, r12, r5<<2 ++ /* r3 = R[j]*(2-R[j]*d) */ ++ mulu.d r4, r3, r5 ++ lsl r3, r5, 2 ++ ++ /* Third approximation */ ++ /* Approximating to 32 bits */ ++ /* r5 = R[j]*d */ ++ mulu.d r4, r3, r9 ++ /* r5 = 2-R[j]*d */ ++ sub r5, r12, r5<<2 ++ /* r3 = R[j]*(2-R[j]*d) */ ++ mulu.d r4, r3, r5 ++ lsl r3, r5, 2 ++ ++ /* Fourth approximation */ ++ /* Approximating to 64 bits */ ++ /* r5,r4 = R[j]*d */ ++ mul_approx_df r3 /*ah*/, r2 /*al*/, r9 /*bh*/, r8 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ ++ lsl r5, 2 ++ or r5, r5, r4>>30 ++ lsl r4, 2 ++ /* r5,r4 = 2-R[j]*d */ ++ neg r4 ++ sbc r5, r12, r5 ++ /* r3,r2 = R[j]*(2-R[j]*d) */ ++ mul_approx_df r3 /*ah*/, r2 /*al*/, r5 /*bh*/, r4 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ ++ lsl r3, r5, 2 ++ or r3, r3, r4>>30 ++ lsl r2, r4, 2 ++ ++ ++ /* Fifth approximation */ ++ /* Approximating to 64 bits */ ++ /* r5,r4 = R[j]*d */ ++ mul_approx_df r3 /*ah*/, r2 /*al*/, r9 /*bh*/, r8 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ ++ lsl r5, 2 ++ or r5, r5, r4>>30 ++ lsl r4, 2 ++ /* r5,r4 = 2-R[j]*d */ ++ neg r4 ++ sbc r5, r12, r5 ++ /* r3,r2 = R[j]*(2-R[j]*d) */ ++ mul_approx_df r3 /*ah*/, r2 /*al*/, r5 /*bh*/, r4 /*bl*/, r5 /*rh*/, r4 /*rl*/, r1 /*sh*/, r0 /*sl*/ ++ lsl r3, r5, 2 ++ or r3, r3, r4>>30 ++ lsl r2, r4, 2 ++ ++ ++ /* Multiply with dividend to get quotient */ ++ mul_approx_df r3 /*ah*/, r2 /*al*/, r11 /*bh*/, r10 /*bl*/, r3 /*rh*/, r2 /*rl*/, r1 /*sh*/, r0 /*sl*/ ++ ++ /* Shift by 3 to get result in 1.63 format, as required by the exponent. */ ++ /* Note that 1.63 format is already used by the exponent in r7, since */ ++ /* a bias of 1023 was added to the result exponent, even though the implicit */ ++ /* bit was inserted. This gives the exponent an additional bias of 1, which */ ++ /* supports 1.63 format. */ ++ lsl r3, r3, 3 ++ or r3, r3, r2>>29 ++ lsl r2, r2, 3 ++ ++ /* To increase speed, this result is not corrected before final rounding.*/ ++ /* This may give a difference to IEEE compliant code of 1 ULP.*/ ++ ++ /* Adjust exponent and mantissa */ ++ /* r7:exp, [r3, r2]:mant, [r12,r11,r10]:scratch*/ ++ /* Mantissa may be of the format 0.xxxx or 1.xxxx. */ ++ /* In the first case, shift one pos to left.*/ ++ sub r10, r7, 1 ++ mov r12, r3 ++ lsl r11, r2, 1 ++ rol r12 ++ bld r3, 31 ++ movne r7, r10 ++ movne r3, r12 ++ movne r2, r11 ++ cp r7, 0 ++ breq 15f /*Result was subnormal. Flush-to-zero and return zero*/ ++ ++ /* Result was not subnormal. Perform rounding. */ ++ /* Note that the tie case (for round-to-even) can not occur in division. */ ++ /* [r7]:exp, [r3, r2]:mant */ ++ /* Mantissa is in 0.64 format. Round by adding 1<<(64-(52+2))=1<<10*/ ++ mov r12, (1<<10) ++ add r2, r12 ++ acr r3 ++ /* Adjust exponent if we overflowed.*/ ++ subcs r7, -1 ++ ++ ++ ++ /* Pack final result*/ ++ /* Input: [r7]:exp, [r3, r2]:mant */ ++ /* Result in [r11,r10] */ ++ /* Insert exponent and sign bit*/ ++ lsl r11, r7, 20 ++ bld lr, 31 ++ bst r11, 31 ++ /* Insert mantissa */ ++ cbr r3, 31 /*Clear implicit bit*/ ++ or r11, r11, r3>>11 ++ lsr r10, r2, 11 ++ or r10, r10, r3<<21 ++ /* Return result in [r11,r10] */ ++ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc ++ ++ ++2: ++ /* Op1 is NaN or inf */ ++ andh r11, 0x000f /* Extract mantissa */ ++ or r11, r10 ++ brne 16f /* Return NaN if op1 is NaN */ ++ /* Op1 is inf check op2 */ ++ lsr r6, r9, 20 /* Extract exponent */ ++ cbr r6, 8 /* Clear sign bit */ ++ cp r6, 0x7ff ++ brne 17f /* Inf/number gives inf, return inf */ ++ rjmp 16f /* The rest gives NaN*/ ++ ++3: ++ /* Op1 is a valid number. Op 2 is NaN or inf */ ++ andh r9, 0x000f /* Extract mantissa */ ++ or r9, r8 ++ brne 16f /* Return NaN if op2 is NaN */ ++ rjmp 15f /* Op2 was inf, return zero*/ ++ ++11: /* Op1 was denormal. Fix it. */ ++ lsl r11, 2 ++ or r11, r11, r10 >> 30 ++ lsl r10, 2 ++ cbr r11, 31 ++ /* Check if op1 is zero. */ ++ or r4, r10, r11 ++ breq __avr32_f64_div_op1_zero ++ normalize_df r7 /*exp*/, r10, r11 /*Mantissa*/, r4, r5 /*scratch*/ ++ lsr r10, 2 ++ or r10, r10, r11 << 30 ++ lsr r11, 2 ++ rjmp 22b ++ ++ ++13: /* Op2 was denormal. Fix it */ ++ lsl r9, 2 ++ or r9, r9, r8 >> 30 ++ lsl r8, 2 ++ cbr r9, 31 ++ /* Check if op2 is zero. */ ++ or r4, r9, r8 ++ breq 17f /* Divisor is zero -> return Inf */ ++ normalize_df r6 /*exp*/, r8, r9 /*Mantissa*/, r4, r5 /*scratch*/ ++ lsr r8, 2 ++ or r8, r8, r9 << 30 ++ lsr r9, 2 ++ rjmp 23b ++ ++ ++15: /* Divide result was subnormal. Return zero. */ ++ mov r11, lr /*Get correct sign*/ ++ andh r11, 0x8000, COH ++ mov r10, 0 ++ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc ++ ++16: /* Return NaN. */ ++ mov r11, -1 ++ mov r10, -1 ++ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc ++ ++17: /* Return INF. */ ++ mov r11, lr /*Get correct sign*/ ++ andh r11, 0x8000, COH ++ orh r11, 0x7ff0 ++ mov r10, 0 ++ ldm sp++, r0, r1, r2, r3, r4, r5, r6, r7,pc ++ ++__avr32_f64_div_op1_zero: ++ or r5, r8, r9 << 1 ++ breq 16b /* 0.0/0.0 -> NaN */ ++ bfextu r4, r9, 20, 11 ++ cp r4, 0x7ff ++ brne 15b /* Return zero */ ++ /* Check if divisor is Inf or NaN */ ++ or r5, r8, r9 << 12 ++ breq 15b /* Divisor is inf -> return zero */ ++ rjmp 16b /* Return NaN */ ++ ++ ++ ++ ++#endif ++ ++#if defined(L_avr32_f32_addsub) || defined(L_avr32_f32_addsub_fast) ++ ++ .align 2 ++__avr32_f32_sub_from_add: ++ /* Switch sign on op2 */ ++ eorh r11, 0x8000 ++ ++#if defined(L_avr32_f32_addsub_fast) ++ .global __avr32_f32_sub_fast ++ .type __avr32_f32_sub_fast,@function ++__avr32_f32_sub_fast: ++#else ++ .global __avr32_f32_sub ++ .type __avr32_f32_sub,@function ++__avr32_f32_sub: ++#endif ++ ++ /* Check signs */ ++ eor r8, r11, r12 ++ /* Different signs, use subtraction. */ ++ brmi __avr32_f32_add_from_sub ++ ++ /* Get sign of op1 */ ++ mov r8, r12 ++ andh r12, 0x8000, COH ++ ++ /* Remove sign from operands */ ++ cbr r11, 31 ++#if defined(L_avr32_f32_addsub_fast) ++ reteq r8 /* If op2 is zero return op1 */ ++#endif ++ cbr r8, 31 ++ ++ /* Put the number with the largest exponent in r10 ++ and the number with the smallest exponent in r9 */ ++ max r10, r8, r11 ++ min r9, r8, r11 ++ cp r10, r8 /*If largest operand (in R10) is not equal to op1*/ ++ subne r12, 1 /* Subtract 1 from sign, which will invert MSB of r12*/ ++ andh r12, 0x8000, COH /*Mask all but MSB*/ ++ ++ /* Unpack exponent and mantissa of op1 */ ++ lsl r8, r10, 8 ++ sbr r8, 31 /* Set implicit bit. */ ++ lsr r10, 23 ++ ++ /* op1 is NaN or Inf. */ ++ cp.w r10, 0xff ++ breq __avr32_f32_sub_op1_nan_or_inf ++ ++ /* Unpack exponent and mantissa of op2 */ ++ lsl r11, r9, 8 ++ sbr r11, 31 /* Set implicit bit. */ ++ lsr r9, 23 ++ ++#if defined(L_avr32_f32_addsub) ++ /* Keep sticky bit for correct IEEE rounding */ ++ st.w --sp, r12 ++ ++ /* op2 is either zero or subnormal. */ ++ breq __avr32_f32_sub_op2_subnormal ++0: ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r9, r10 ++ ++ /* Saturate the shift amount to 31. If the amount ++ is any larger op2 is insignificant. */ ++ satu r9 >> 0, 5 ++ ++ /* Shift mantissa of op2 to same decimal point as the mantissa ++ of op1. */ ++ lsr r12, r11, r9 ++ ++ /* Put the remainding bits into r11[23:..].*/ ++ rsub r9, r9, (32-8) ++ lsl r11, r11, r9 ++ ++ /* Now subtract the mantissas. */ ++ sub r8, r12 ++ ++ ld.w r12, sp++ ++ ++ /* Normalize resulting mantissa. */ ++ clz r9, r8 ++ lsl r8, r8, r9 ++ sub r10, r9 ++ brle __avr32_f32_sub_subnormal_result ++ ++ /* Insert the bits we will remove from the mantissa into r11[31:24] */ ++ bfins r11, r8, 24, 8 ++#else ++ /* Ignore sticky bit to simplify and speed up rounding */ ++ /* op2 is either zero or subnormal. */ ++ breq __avr32_f32_sub_op2_subnormal ++0: ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r9, r10 ++ ++ /* Saturate the shift amount to 31. If the amount ++ is any larger op2 is insignificant. */ ++ satu r9 >> 0, 5 ++ ++ /* Shift mantissa of op2 to same decimal point as the mantissa ++ of op1. */ ++ lsr r11, r11, r9 ++ ++ /* Now subtract the mantissas. */ ++ sub r8, r11 ++ ++ /* Normalize resulting mantissa. */ ++ clz r9, r8 ++ lsl r8, r8, r9 ++ sub r10, r9 ++ brle __avr32_f32_sub_subnormal_result ++#endif ++ ++ /* Pack result. */ ++ or r12, r12, r8 >> 8 ++ bfins r12, r10, 23, 8 ++ ++ /* Round */ ++__avr32_f32_sub_round: ++#if defined(L_avr32_f32_addsub) ++ mov r10, 0 ++ sbr r10, 31 ++ bld r12, 0 ++ subne r10, -1 ++ cp.w r11, r10 ++ subhs r12, -1 ++#else ++ bld r8, 7 ++ acr r12 ++#endif ++ ++ ret r12 ++ ++ ++__avr32_f32_sub_op2_subnormal: ++ /* Fix implicit bit and adjust exponent of subnormals. */ ++ cbr r11, 31 ++ /* Set exponent to 1 if we do not have a zero. */ ++ movne r9,1 ++ ++ /* Check if op1 is also subnormal. */ ++ cp.w r10, 0 ++ brne 0b ++ ++ cbr r8, 31 ++ /* If op1 is not zero set exponent to 1. */ ++ movne r10,1 ++ ++ rjmp 0b ++ ++__avr32_f32_sub_op1_nan_or_inf: ++ /* Check if op1 is NaN, if so return NaN */ ++ lsl r11, r8, 1 ++ retne -1 ++ ++ /* op1 is Inf. */ ++ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ ++ ++ /* Check if op2 is Inf. or NaN */ ++ lsr r11, r9, 23 ++ cp.w r11, 0xff ++ retne r12 /* op2 not Inf or NaN, return op1 */ ++ ++ ret -1 /* op2 Inf or NaN, return NaN */ ++ ++__avr32_f32_sub_subnormal_result: ++ /* Check if the number is so small that ++ it will be represented with zero. */ ++ rsub r10, r10, 9 ++ rsub r9, r10, 32 ++ retcs 0 ++ ++ /* Shift the mantissa into the correct position.*/ ++ lsr r10, r8, r10 ++ /* Add sign bit. */ ++ or r10, r12 ++ ++ /* Put the shifted out bits in the most significant part ++ of r8. */ ++ lsl r8, r8, r9 ++ ++#if defined(L_avr32_f32_addsub) ++ /* Add all the remainder bits used for rounding into r11 */ ++ andh r11, 0x00FF ++ or r11, r8 ++#else ++ lsr r8, 24 ++#endif ++ rjmp __avr32_f32_sub_round ++ ++ ++ .align 2 ++ ++__avr32_f32_add_from_sub: ++ /* Switch sign on op2 */ ++ eorh r11, 0x8000 ++ ++#if defined(L_avr32_f32_addsub_fast) ++ .global __avr32_f32_add_fast ++ .type __avr32_f32_add_fast,@function ++__avr32_f32_add_fast: ++#else ++ .global __avr32_f32_add ++ .type __avr32_f32_add,@function ++__avr32_f32_add: ++#endif ++ ++ /* Check signs */ ++ eor r8, r11, r12 ++ /* Different signs, use subtraction. */ ++ brmi __avr32_f32_sub_from_add ++ ++ /* Get sign of op1 */ ++ mov r8, r12 ++ andh r12, 0x8000, COH ++ ++ /* Remove sign from operands */ ++ cbr r11, 31 ++#if defined(L_avr32_f32_addsub_fast) ++ reteq r8 /* If op2 is zero return op1 */ ++#endif ++ cbr r8, 31 ++ ++ /* Put the number with the largest exponent in r10 ++ and the number with the smallest exponent in r9 */ ++ max r10, r8, r11 ++ min r9, r8, r11 ++ ++ /* Unpack exponent and mantissa of op1 */ ++ lsl r8, r10, 8 ++ sbr r8, 31 /* Set implicit bit. */ ++ lsr r10, 23 ++ ++ /* op1 is NaN or Inf. */ ++ cp.w r10, 0xff ++ breq __avr32_f32_add_op1_nan_or_inf ++ ++ /* Unpack exponent and mantissa of op2 */ ++ lsl r11, r9, 8 ++ sbr r11, 31 /* Set implicit bit. */ ++ lsr r9, 23 ++ ++#if defined(L_avr32_f32_addsub) ++ /* Keep sticky bit for correct IEEE rounding */ ++ st.w --sp, r12 ++ ++ /* op2 is either zero or subnormal. */ ++ breq __avr32_f32_add_op2_subnormal ++0: ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r9, r10 ++ ++ /* Saturate the shift amount to 31. If the amount ++ is any larger op2 is insignificant. */ ++ satu r9 >> 0, 5 ++ ++ /* Shift mantissa of op2 to same decimal point as the mantissa ++ of op1. */ ++ lsr r12, r11, r9 ++ ++ /* Put the remainding bits into r11[23:..].*/ ++ rsub r9, r9, (32-8) ++ lsl r11, r11, r9 ++ /* Insert the bits we will remove from the mantissa into r11[31:24] */ ++ bfins r11, r12, 24, 8 ++ ++ /* Now add the mantissas. */ ++ add r8, r12 ++ ++ ld.w r12, sp++ ++#else ++ /* Ignore sticky bit to simplify and speed up rounding */ ++ /* op2 is either zero or subnormal. */ ++ breq __avr32_f32_add_op2_subnormal ++0: ++ /* Get shift amount to scale mantissa of op2. */ ++ rsub r9, r10 ++ ++ /* Saturate the shift amount to 31. If the amount ++ is any larger op2 is insignificant. */ ++ satu r9 >> 0, 5 ++ ++ /* Shift mantissa of op2 to same decimal point as the mantissa ++ of op1. */ ++ lsr r11, r11, r9 ++ ++ /* Now add the mantissas. */ ++ add r8, r11 ++ ++#endif ++ /* Check if we overflowed. */ ++ brcs __avr32_f32_add_res_of ++1: ++ /* Pack result. */ ++ or r12, r12, r8 >> 8 ++ bfins r12, r10, 23, 8 ++ ++ /* Round */ ++#if defined(L_avr32_f32_addsub) ++ mov r10, 0 ++ sbr r10, 31 ++ bld r12, 0 ++ subne r10, -1 ++ cp.w r11, r10 ++ subhs r12, -1 ++#else ++ bld r8, 7 ++ acr r12 ++#endif ++ ++ ret r12 ++ ++__avr32_f32_add_op2_subnormal: ++ /* Fix implicit bit and adjust exponent of subnormals. */ ++ cbr r11, 31 ++ /* Set exponent to 1 if we do not have a zero. */ ++ movne r9,1 ++ ++ /* Check if op1 is also subnormal. */ ++ cp.w r10, 0 ++ brne 0b ++ ++ cbr r8, 31 ++ /* If op1 is not zero set exponent to 1. */ ++ movne r10,1 ++ ++ rjmp 0b ++ ++__avr32_f32_add_op1_nan_or_inf: ++ /* Check if op1 is NaN, if so return NaN */ ++ lsl r11, r8, 1 ++ retne -1 ++ ++ /* op1 is Inf. */ ++ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ ++ ++ /* Check if op2 is Inf. or NaN */ ++ lsr r11, r9, 23 ++ cp.w r11, 0xff ++ retne r12 /* op2 not Inf or NaN, return op1 */ ++ ++ lsl r9, 9 ++ reteq r12 /* op2 Inf return op1 */ ++ ret -1 /* op2 is NaN, return NaN */ ++ ++__avr32_f32_add_res_of: ++ /* We overflowed. Increase exponent and shift mantissa.*/ ++ lsr r8, 1 ++ sub r10, -1 ++ ++ /* Clear mantissa to set result to Inf if the exponent is 255. */ ++ cp.w r10, 255 ++ moveq r8, 0 ++ moveq r11, 0 ++ rjmp 1b ++ ++ ++#endif ++ ++ ++#if defined(L_avr32_f32_div) || defined(L_avr32_f32_div_fast) ++ .align 2 ++ ++#if defined(L_avr32_f32_div_fast) ++ .global __avr32_f32_div_fast ++ .type __avr32_f32_div_fast,@function ++__avr32_f32_div_fast: ++#else ++ .global __avr32_f32_div ++ .type __avr32_f32_div,@function ++__avr32_f32_div: ++#endif ++ ++ eor r8, r11, r12 /* MSB(r8) = Sign(op1) ^ Sign(op2) */ ++ ++ /* Unpack */ ++ lsl r12,1 ++ reteq 0 /* Return zero if op1 is zero */ ++ lsl r11,1 ++ breq 4f /* Check op2 for zero */ ++ ++ /* Unpack op1*/ ++ /* exp: r9 */ ++ /* sf: r12 */ ++ lsr r9, r12, 24 ++ breq 11f /*If number is subnormal*/ ++ cp r9, 0xff ++ brhs 2f /* Check op1 for NaN or Inf */ ++ lsl r12, 7 ++ sbr r12, 31 /*Implicit bit*/ ++12: ++ ++ /* Unpack op2*/ ++ /* exp: r10 */ ++ /* sf: r11 */ ++ lsr r10, r11, 24 ++ breq 13f /*If number is subnormal*/ ++ cp r10, 0xff ++ brhs 3f /* Check op2 for NaN or Inf */ ++ ++ lsl r11,7 ++ sbr r11, 31 /*Implicit bit*/ ++14: ++ ++ /* For UC3, store with predecrement is faster than stm */ ++ st.w --sp, r5 ++ st.d --sp, r6 ++ ++ /* Calculate new exponent */ ++ sub r9, r10 ++ sub r9,-127 ++ ++ /* Divide */ ++ /* Approximating 1/d with the following recurrence: */ ++ /* R[j+1] = R[j]*(2-R[j]*d) */ ++ /* Using 2.30 format */ ++ /* TWO: r10 */ ++ /* d: r5 */ ++ /* Multiply result : r6, r7 */ ++ /* Initial guess : r11 */ ++ /* New approximations : r11 */ ++ /* Dividend : r12 */ ++ ++ mov r10, 1 /* Load TWO */ ++ brev r10 ++ ++ lsr r12, 2 /* Get significand of Op1 in 2.30 format */ ++ lsr r5, r11, 2 /* Get significand of Op2 (=d) in 2.30 format */ ++ ++ /* Load initial guess, using look-up table */ ++ /* Initial guess is of format 01.XY, where XY is constructed as follows: */ ++ /* Let d be of following format: 00.1xy....., then XY=~xy */ ++ /* For d=00.100 = 0,5 -> initial guess=01.11 = 1,75 */ ++ /* For d=00.101 = 0,625 -> initial guess=01.11 = 1,5 */ ++ /* For d=00.110 = 0,75 -> initial guess=01.11 = 1,25 */ ++ /* For d=00.111 = 0,875 -> initial guess=01.11 = 1,0 */ ++ ++ lsr r11, r10, 1 ++ bfextu r6, r5, 27, 2 ++ com r6 ++ bfins r11, r6, 28, 2 ++ ++ /* First approximation */ ++ /* r7 = R[j]*d */ ++ mulu.d r6, r11, r5 ++ /* r7 = 2-R[j]*d */ ++ sub r7, r10, r7<<2 ++ /* r11 = R[j]*(2-R[j]*d) */ ++ mulu.d r6, r11, r7 ++ lsl r11, r7, 2 ++ ++ /* Second approximation */ ++ /* r7 = R[j]*d */ ++ mulu.d r6, r11, r5 ++ /* r7 = 2-R[j]*d */ ++ sub r7, r10, r7<<2 ++ /* r11 = R[j]*(2-R[j]*d) */ ++ mulu.d r6, r11, r7 ++ lsl r11, r7, 2 ++ ++ /* Third approximation */ ++ /* r7 = R[j]*d */ ++ mulu.d r6, r11, r5 ++ /* r7 = 2-R[j]*d */ ++ sub r7, r10, r7<<2 ++ /* r11 = R[j]*(2-R[j]*d) */ ++ mulu.d r6, r11, r7 ++ lsl r11, r7, 2 ++ ++ /* Fourth approximation */ ++ /* r7 = R[j]*d */ ++ mulu.d r6, r11, r5 ++ /* r7 = 2-R[j]*d */ ++ sub r7, r10, r7<<2 ++ /* r11 = R[j]*(2-R[j]*d) */ ++ mulu.d r6, r11, r7 ++ lsl r11, r7, 2 ++ ++ ++ /* Multiply with dividend to get quotient, r7 = sf(op1)/sf(op2) */ ++ mulu.d r6, r11, r12 ++ ++ /* Shift by 3 to get result in 1.31 format, as required by the exponent. */ ++ /* Note that 1.31 format is already used by the exponent in r9, since */ ++ /* a bias of 127 was added to the result exponent, even though the implicit */ ++ /* bit was inserted. This gives the exponent an additional bias of 1, which */ ++ /* supports 1.31 format. */ ++ lsl r10, r7, 3 ++ ++#if defined(L_avr32_f32_div) ++ /* To perform correct rounding, check for nonzero remainder, */ ++ /* and set LSB in quot if remainder != 0 */ ++ /* Remainder = dividend(r12) - divisor(r5)*quotient(r10) */ ++ ++ lsl r5, 1 /* Transform divisor from 2.30 to 1.31 format */ ++ /* Mask all bits lower than guard in quotient. */ ++ /* These bits are inexact due to approximative algorithm */ ++ andl r10, 0xffc0 ++ /* Add 1 in least significant bit pos to make sure approximation is from above */ ++ sub r10, -64 ++ mulu.d r6, r5, r10 ++ /* If remainder < 0, truncated quotient is too large, so the */ ++ /* delta added must be subtracted to get the correct truncated quotient. */ ++ sub r12, r7 /* Calculate remainder and implicitly set flags */ ++ sublt r10, 64 ++#endif ++ ++ /* For UC3, load with postincrement is faster than ldm */ ++ ld.d r6, sp++ ++ ld.w r5, sp++ ++ ++ /* Adjust exponent and mantissa */ ++ /* r9:exp, r10:mant, r11:scratch*/ ++ clz r11, r10 ++ sub r9, r11 ++ breq 16f /*Result was subnormal*/ ++ lsl r10, r10, r11 ++ ++ /* Result was not subnormal. Perform rounding. */ ++ /* Note that the tie case (for round-to-even) can not occur in division. */ ++ /* r9:exp, r10:mant*/ ++ sub r10, -1*(0x80) ++ /* Adjust exponent if we overflowed. Note that we must use {cc} ++ since we perform the add using a sub insn. */ ++ subcc r9, -1 ++ ++ /* Pack final result*/ ++ lsr r12, r10, 7 ++ bfins r12, r9, 24, 8 ++__divsf_return_op1: ++ lsl r8, 1 ++ ror r12 ++ ret r12 ++ ++ ++2: ++ /* Op1 is NaN or inf */ ++ retne -1 /* Return NaN if op1 is NaN */ ++ /* Op1 is inf check op2 */ ++ mov r9, 0xff ++ brev r9 ++ cp r11, r9 ++ brlo __divsf_return_op1 /* inf/number gives inf */ ++ ret -1 /* The rest gives NaN*/ ++3: ++ /* Op1 is NaN or inf */ ++ reteq 0 /* Return zero if number/inf*/ ++ ret -1 /* Return NaN*/ ++4: ++ /* Op2 is zero ? */ ++ tst r12,r12 ++ reteq -1 /* 0.0/0.0 is NaN */ ++ /* Nonzero/0.0 is Inf. Sign bit will be shifted in before returning*/ ++ mov r12, 0x7ff ++ brev r12 ++ rjmp __divsf_return_op1 ++ ++11: /* Op1 was denormal. Fix it. */ ++ lsl r12,7 ++ clz r9,r12 ++ lsl r12,r12,r9 ++ rsub r9,r9,1 ++ rjmp 12b ++ ++13: /* Op2 was denormal. Fix it. */ ++ lsl r11,7 ++ clz r10,r11 ++ lsl r11,r11,r10 ++ rsub r10,r10,1 ++ rjmp 14b ++ ++ ++16: /* Divide result was subnormal. Fix it and return. */ ++#if defined(L_avr32_f32_div) ++ lsl r10, r10, r11 /*Perform shift required by adjustment of exponent and mantissa*/ ++ adjust_subnormal_sf r12 /*sf*/, r9 /*exp*/, r10 /*mant*/, r8 /*sign*/,r11 /*scratch*/ ++#else ++ /*Flush to zero*/ ++ mov r12, 0 ++#endif ++ ret r12 ++ ++#endif ++ ++#ifdef L_avr32_f32_mul ++ .global __avr32_f32_mul ++ .type __avr32_f32_mul,@function ++ ++ ++__avr32_f32_mul: ++ mov r8, r12 ++ eor r12, r11 /* MSB(r8) = Sign(op1) ^ Sign(op2) */ ++ andh r12, 0x8000, COH ++ ++ /* arrange operands so that that op1 >= op2 */ ++ cbr r8, 31 ++ breq __avr32_f32_mul_op1_zero ++ cbr r11, 31 ++ ++ /* Put the number with the largest exponent in r10 ++ and the number with the smallest exponent in r9 */ ++ max r10, r8, r11 ++ min r9, r8, r11 ++ ++ /* Unpack exponent and mantissa of op1 */ ++ lsl r8, r10, 8 ++ sbr r8, 31 /* Set implicit bit. */ ++ lsr r10, 23 ++ ++ /* op1 is NaN or Inf. */ ++ cp.w r10, 0xff ++ breq __avr32_f32_mul_op1_nan_or_inf ++ ++ /* Unpack exponent and mantissa of op2 */ ++ lsl r11, r9, 8 ++ sbr r11, 31 /* Set implicit bit. */ ++ lsr r9, 23 ++ ++ /* op2 is either zero or subnormal. */ ++ breq __avr32_f32_mul_op2_subnormal ++0: ++ /* Calculate new exponent */ ++ add r9,r10 ++ ++ /* Do the multiplication */ ++ mulu.d r10,r8,r11 ++ ++ /* We might need to scale up by two if the MSB of the result is ++ zero. */ ++ lsl r8, r11, 1 ++ movcc r11, r8 ++ subcc r9, 1 ++ ++ /* Put the shifted out bits of the mantissa into r10 */ ++ lsr r10, 8 ++ bfins r10, r11, 24, 8 ++ ++ sub r9,(127-1) /* remove extra exponent bias */ ++ brle __avr32_f32_mul_res_subnormal ++ ++ /* Check for Inf. */ ++ cp.w r9, 0xff ++ brge 1f ++ ++ /* Pack result. */ ++ or r12, r12, r11 >> 8 ++ bfins r12, r9, 23, 8 ++ ++ /* Round */ ++__avr32_f32_mul_round: ++ mov r8, 0 ++ sbr r8, 31 ++ bld r12, 0 ++ subne r8, -1 ++ ++ cp.w r10, r8 ++ subhs r12, -1 ++ ++ ret r12 ++ ++1: ++ /* Return Inf */ ++ orh r12, 0x7f80 ++ ret r12 ++ ++__avr32_f32_mul_op2_subnormal: ++ cbr r11, 31 ++ clz r9, r11 ++ retcs 0 /* op2 is zero. Return 0 */ ++ lsl r11, r11, r9 ++ rsub r9, r9, 1 ++ ++ /* Check if op2 is subnormal. */ ++ tst r10, r10 ++ brne 0b ++ ++ /* op2 is subnormal */ ++ cbr r8, 31 ++ clz r10, r11 ++ retcs 0 /* op1 is zero. Return 0 */ ++ lsl r8, r8, r10 ++ rsub r10, r10, 1 ++ ++ rjmp 0b ++ ++ ++__avr32_f32_mul_op1_nan_or_inf: ++ /* Check if op1 is NaN, if so return NaN */ ++ lsl r11, r8, 1 ++ retne -1 ++ ++ /* op1 is Inf. */ ++ tst r9, r9 ++ reteq -1 /* Inf * 0 -> NaN */ ++ ++ bfins r12, r10, 23, 8 /* Generate Inf in r12 */ ++ ++ /* Check if op2 is Inf. or NaN */ ++ lsr r11, r9, 23 ++ cp.w r11, 0xff ++ retne r12 /* op2 not Inf or NaN, return Info */ ++ ++ lsl r9, 9 ++ reteq r12 /* op2 Inf return Inf */ ++ ret -1 /* op2 is NaN, return NaN */ ++ ++__avr32_f32_mul_res_subnormal: ++ /* Check if the number is so small that ++ it will be represented with zero. */ ++ rsub r9, r9, 9 ++ rsub r8, r9, 32 ++ retcs 0 ++ ++ /* Shift the mantissa into the correct position.*/ ++ lsr r9, r11, r9 ++ /* Add sign bit. */ ++ or r12, r9 ++ /* Put the shifted out bits in the most significant part ++ of r8. */ ++ lsl r11, r11, r8 ++ ++ /* Add all the remainder bits used for rounding into r11 */ ++ andh r10, 0x00FF ++ or r10, r11 ++ rjmp __avr32_f32_mul_round ++ ++__avr32_f32_mul_op1_zero: ++ bfextu r10, r11, 23, 8 ++ cp.w r10, 0xff ++ retne r12 ++ reteq -1 ++ ++#endif ++ ++ ++#ifdef L_avr32_s32_to_f32 ++ .global __avr32_s32_to_f32 ++ .type __avr32_s32_to_f32,@function ++__avr32_s32_to_f32: ++ cp r12, 0 ++ reteq r12 /* If zero then return zero float */ ++ mov r11, r12 /* Keep the sign */ ++ abs r12 /* Compute the absolute value */ ++ mov r10, 31 + 127 /* Set the correct exponent */ ++ ++ /* Normalize */ ++ normalize_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ ++ ++ /* Check for subnormal result */ ++ cp.w r10, 0 ++ brle __avr32_s32_to_f32_subnormal ++ ++ round_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ ++ pack_sf r12 /*sf*/, r10 /*exp*/, r12 /*mant*/ ++ lsl r11, 1 ++ ror r12 ++ ret r12 ++ ++__avr32_s32_to_f32_subnormal: ++ /* Adjust a subnormal result */ ++ adjust_subnormal_sf r12/*sf*/, r10 /*exp*/, r12 /*mant*/, r11/*sign*/, r9 /*scratch*/ ++ ret r12 ++ ++#endif ++ ++#ifdef L_avr32_u32_to_f32 ++ .global __avr32_u32_to_f32 ++ .type __avr32_u32_to_f32,@function ++__avr32_u32_to_f32: ++ cp r12, 0 ++ reteq r12 /* If zero then return zero float */ ++ mov r10, 31 + 127 /* Set the correct exponent */ ++ ++ /* Normalize */ ++ normalize_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ ++ ++ /* Check for subnormal result */ ++ cp.w r10, 0 ++ brle __avr32_u32_to_f32_subnormal ++ ++ round_sf r10 /*exp*/, r12 /*mant*/, r9 /*scratch*/ ++ pack_sf r12 /*sf*/, r10 /*exp*/, r12 /*mant*/ ++ lsr r12,1 /* Sign bit is 0 for unsigned int */ ++ ret r12 ++ ++__avr32_u32_to_f32_subnormal: ++ /* Adjust a subnormal result */ ++ mov r8, 0 ++ adjust_subnormal_sf r12/*sf*/,r10 /*exp*/, r12 /*mant*/,r8/*sign*/, r9 /*scratch*/ ++ ret r12 ++ ++ ++#endif ++ ++ ++#ifdef L_avr32_f32_to_s32 ++ .global __avr32_f32_to_s32 ++ .type __avr32_f32_to_s32,@function ++__avr32_f32_to_s32: ++ bfextu r11, r12, 23, 8 ++ sub r11,127 /* Fix bias */ ++ retlo 0 /* Negative exponent yields zero integer */ ++ ++ /* Shift mantissa into correct position */ ++ rsub r11,r11,31 /* Shift amount */ ++ lsl r10,r12,8 /* Get mantissa */ ++ sbr r10,31 /* Add implicit bit */ ++ lsr r10,r10,r11 /* Perform shift */ ++ lsl r12,1 /* Check sign */ ++ retcc r10 /* if positive, we are done */ ++ neg r10 /* if negative float, negate result */ ++ ret r10 ++ ++#endif ++ ++#ifdef L_avr32_f32_to_u32 ++ .global __avr32_f32_to_u32 ++ .type __avr32_f32_to_u32,@function ++__avr32_f32_to_u32: ++ cp r12,0 ++ retmi 0 /* Negative numbers gives 0 */ ++ bfextu r11, r12, 23, 8 /* Extract exponent */ ++ sub r11,127 /* Fix bias */ ++ retlo 0 /* Negative exponent yields zero integer */ ++ ++ /* Shift mantissa into correct position */ ++ rsub r11,r11,31 /* Shift amount */ ++ lsl r12,8 /* Get mantissa */ ++ sbr r12,31 /* Add implicit bit */ ++ lsr r12,r12,r11 /* Perform shift */ ++ ret r12 ++ ++#endif ++ ++#ifdef L_avr32_f32_to_f64 ++ .global __avr32_f32_to_f64 ++ .type __avr32_f32_to_f64,@function ++ ++__avr32_f32_to_f64: ++ lsl r11,r12,1 /* Remove sign bit, keep original value in r12*/ ++ moveq r10, 0 ++ reteq r11 /* Return zero if input is zero */ ++ ++ bfextu r9,r11,24,8 /* Get exponent */ ++ cp.w r9,0xff /* check for NaN or inf */ ++ breq 0f ++ ++ lsl r11,7 /* Convert sf mantissa to df format */ ++ mov r10,0 ++ ++ /* Check if implicit bit should be set */ ++ cp.w r9, 0 ++ subeq r9,-1 /* Adjust exponent if it was 0 */ ++ srne r8 ++ or r11, r11, r8 << 31 /* Set implicit bit if needed */ ++ sub r9,(127-0x3ff) /* Convert exponent to df format exponent */ ++ ++ /*We know that low register of mantissa is 0, and will be unaffected by normalization.*/ ++ /*We can therefore use the faster normalize_sf function instead of normalize_df.*/ ++ normalize_sf r9 /*exp*/, r11 /*mantissa*/, r8 /*scratch*/ ++ pack_df r9 /*exp*/, r10, r11 /*mantissa*/, r10, r11 /*df*/ ++ ++__extendsfdf_return_op1: ++ /* Rotate in sign bit */ ++ lsl r12, 1 ++ ror r11 ++ ret r11 ++ ++0: ++ /* Inf or NaN*/ ++ mov r10, 0 ++ orh r10, 0xffe0 ++ lsl r11,8 /* check mantissa */ ++ movne r11, -1 /* Return NaN */ ++ moveq r11, r10 /* Return inf */ ++ rjmp __extendsfdf_return_op1 ++#endif ++ ++ ++#ifdef L_avr32_f64_to_f32 ++ .global __avr32_f64_to_f32 ++ .type __avr32_f64_to_f32,@function ++ ++__avr32_f64_to_f32: ++ /* Unpack */ ++ lsl r9,r11,1 /* Unpack exponent */ ++ lsr r9,21 ++ ++ reteq 0 /* If exponent is 0 the number is so small ++ that the conversion to single float gives ++ zero */ ++ ++ lsl r8,r11,10 /* Adjust mantissa */ ++ or r12,r8,r10>>22 ++ ++ lsl r10,10 /* Check if there are any remaining bits ++ in the low part of the mantissa.*/ ++ neg r10 ++ rol r12 /* If there were remaining bits then set lsb ++ of mantissa to 1 */ ++ ++ cp r9,0x7ff ++ breq 2f /* Check for NaN or inf */ ++ ++ sub r9,(0x3ff-127) /* Adjust bias of exponent */ ++ sbr r12,31 /* set the implicit bit.*/ ++ ++ cp.w r9, 0 /* Check for subnormal number */ ++ brle 3f ++ ++ round_sf r9 /*exp*/, r12 /*mant*/, r10 /*scratch*/ ++ pack_sf r12 /*sf*/, r9 /*exp*/, r12 /*mant*/ ++__truncdfsf_return_op1: ++ /* Rotate in sign bit */ ++ lsl r11, 1 ++ ror r12 ++ ret r12 ++ ++2: ++ /* NaN or inf */ ++ cbr r12,31 /* clear implicit bit */ ++ retne -1 /* Return NaN if mantissa not zero */ ++ mov r12, 0 ++ orh r12, 0xff00 ++ ret r12 /* Return inf */ ++ ++3: /* Result is subnormal. Adjust it.*/ ++ adjust_subnormal_sf r12/*sf*/,r9 /*exp*/, r12 /*mant*/, r11/*sign*/, r10 /*scratch*/ ++ ret r12 ++ ++ ++#endif ++ +\ No newline at end of file +diff -Nrup gcc-4.2.1/gcc/config/avr32/lib2funcs.S gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib2funcs.S +--- gcc-4.2.1/gcc/config/avr32/lib2funcs.S 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/lib2funcs.S 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,21 @@ ++ .align 4 ++ .global __nonlocal_goto ++ .type __nonlocal_goto,@function ++ ++/* __nonlocal_goto: This function handles nonlocal_goto's in gcc. ++ ++ parameter 0 (r12) = New Frame Pointer ++ parameter 1 (r11) = Address to goto ++ parameter 2 (r10) = New Stack Pointer ++ ++ This function invalidates the return stack, since it returns from a ++ function without using a return instruction. ++*/ ++__nonlocal_goto: ++ mov r7, r12 ++ mov sp, r10 ++ frs # Flush return stack ++ mov pc, r11 ++ ++ ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/linux-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/linux-elf.h +--- gcc-4.2.1/gcc/config/avr32/linux-elf.h 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/linux-elf.h 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,156 @@ ++/* ++ Linux/Elf specific definitions. ++ Copyright 2003-2006 Atmel Corporation. ++ ++ Written by Ronny Pedersen, Atmel Norway, ++ and H�vard Skinnemoen, Atmel Norway, ++ ++ This file is part of GCC. ++ ++ This program is free software; you can redistribute it and/or modify ++ it under the terms of the GNU General Public License as published by ++ the Free Software Foundation; either version 2 of the License, or ++ (at your option) any later version. ++ ++ This program is distributed in the hope that it will be useful, ++ but WITHOUT ANY WARRANTY; without even the implied warranty of ++ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ GNU General Public License for more details. ++ ++ You should have received a copy of the GNU General Public License ++ along with this program; if not, write to the Free Software ++ Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ ++ ++ ++ ++/* elfos.h should have already been included. Now just override ++ any conflicting definitions and add any extras. */ ++ ++/* Run-time Target Specification. */ ++#undef TARGET_VERSION ++#define TARGET_VERSION fputs (" (AVR32 GNU/Linux with ELF)", stderr); ++ ++/* Do not assume anything about header files. */ ++#define NO_IMPLICIT_EXTERN_C ++ ++/* The GNU C++ standard library requires that these macros be defined. */ ++#undef CPLUSPLUS_CPP_SPEC ++#define CPLUSPLUS_CPP_SPEC "-D_GNU_SOURCE %(cpp)" ++ ++/* Now we define the strings used to build the spec file. */ ++#undef LIB_SPEC ++#define LIB_SPEC \ ++ "%{pthread:-lpthread} \ ++ %{shared:-lc} \ ++ %{!shared:%{profile:-lc_p}%{!profile:-lc}}" ++ ++/* Provide a STARTFILE_SPEC appropriate for GNU/Linux. Here we add ++ the GNU/Linux magical crtbegin.o file (see crtstuff.c) which ++ provides part of the support for getting C++ file-scope static ++ object constructed before entering `main'. */ ++ ++#undef STARTFILE_SPEC ++#define STARTFILE_SPEC \ ++ "%{!shared: \ ++ %{pg:gcrt1.o%s} %{!pg:%{p:gcrt1.o%s} \ ++ %{!p:%{profile:gcrt1.o%s} \ ++ %{!profile:crt1.o%s}}}} \ ++ crti.o%s %{!shared:crtbegin.o%s} %{shared:crtbeginS.o%s}" ++ ++/* Provide a ENDFILE_SPEC appropriate for GNU/Linux. Here we tack on ++ the GNU/Linux magical crtend.o file (see crtstuff.c) which ++ provides part of the support for getting C++ file-scope static ++ object constructed before entering `main', followed by a normal ++ GNU/Linux "finalizer" file, `crtn.o'. */ ++ ++#undef ENDFILE_SPEC ++#define ENDFILE_SPEC \ ++ "%{!shared:crtend.o%s} %{shared:crtendS.o%s} crtn.o%s" ++ ++#undef ASM_SPEC ++#define ASM_SPEC "%{!mno-pic:--pic} %{mrelax|O*:%{mno-relax|O0|O1: ;:--linkrelax}} %{mcpu=*:-mcpu=%*}" ++ ++#undef LINK_SPEC ++#define LINK_SPEC "%{version:-v} \ ++ %{static:-Bstatic} \ ++ %{shared:-shared} \ ++ %{symbolic:-Bsymbolic} \ ++ %{rdynamic:-export-dynamic} \ ++ %{!dynamic-linker:-dynamic-linker /lib/ld-uClibc.so.0} \ ++ %{mrelax|O*:%{mno-relax|O0|O1: ;:--relax}}" ++ ++#define TARGET_OS_CPP_BUILTINS() LINUX_TARGET_OS_CPP_BUILTINS() ++ ++/* This is how we tell the assembler that two symbols have the same value. */ ++#define ASM_OUTPUT_DEF(FILE, NAME1, NAME2) \ ++ do \ ++ { \ ++ assemble_name (FILE, NAME1); \ ++ fputs (" = ", FILE); \ ++ assemble_name (FILE, NAME2); \ ++ fputc ('\n', FILE); \ ++ } \ ++ while (0) ++ ++ ++ ++#undef CC1_SPEC ++#define CC1_SPEC "%{profile:-p}" ++ ++/* Target CPU builtins. */ ++#define TARGET_CPU_CPP_BUILTINS() \ ++ do \ ++ { \ ++ builtin_define ("__avr32__"); \ ++ builtin_define ("__AVR32__"); \ ++ builtin_define ("__AVR32_LINUX__"); \ ++ builtin_define (avr32_part->macro); \ ++ builtin_define (avr32_arch->macro); \ ++ if (avr32_arch->uarch_type == UARCH_TYPE_AVR32A) \ ++ builtin_define ("__AVR32_AVR32A__"); \ ++ else \ ++ builtin_define ("__AVR32_AVR32B__"); \ ++ if (TARGET_UNALIGNED_WORD) \ ++ builtin_define ("__AVR32_HAS_UNALIGNED_WORD__"); \ ++ if (TARGET_SIMD) \ ++ builtin_define ("__AVR32_HAS_SIMD__"); \ ++ if (TARGET_DSP) \ ++ builtin_define ("__AVR32_HAS_DSP__"); \ ++ if (TARGET_RMW) \ ++ builtin_define ("__AVR32_HAS_RMW__"); \ ++ if (TARGET_BRANCH_PRED) \ ++ builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ ++ if (TARGET_FAST_FLOAT) \ ++ builtin_define ("__AVR32_FAST_FLOAT__"); \ ++ if (flag_pic) \ ++ { \ ++ builtin_define ("__PIC__"); \ ++ builtin_define ("__pic__"); \ ++ } \ ++ } \ ++ while (0) ++ ++ ++ ++/* Call the function profiler with a given profile label. */ ++#undef FUNCTION_PROFILER ++#define FUNCTION_PROFILER(STREAM, LABELNO) \ ++ do \ ++ { \ ++ fprintf (STREAM, "\tmov\tlr, lo(mcount)\n\torh\tlr, hi(mcount)\n"); \ ++ fprintf (STREAM, "\ticall lr\n"); \ ++ } \ ++ while (0) ++ ++#define NO_PROFILE_COUNTERS 1 ++ ++/* For dynamic libraries to work */ ++/* #define PLT_REG_CALL_CLOBBERED 1 */ ++#define AVR32_ALWAYS_PIC 1 ++ ++/* uclibc does not implement sinf, cosf etc. */ ++#undef TARGET_C99_FUNCTIONS ++#define TARGET_C99_FUNCTIONS 0 ++ ++#define LINK_GCC_C_SEQUENCE_SPEC \ ++ "%{static:--start-group} %G %L %{static:--end-group}%{!static:%G}" +diff -Nrup gcc-4.2.1/gcc/config/avr32/predicates.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/predicates.md +--- gcc-4.2.1/gcc/config/avr32/predicates.md 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/predicates.md 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,331 @@ ++;; AVR32 predicates file. ++;; Copyright 2003-2006 Atmel Corporation. ++;; ++;; Written by Ronny Pedersen, Atmel Norway, ++;; ++;; This file is part of GCC. ++;; ++;; This program is free software; you can redistribute it and/or modify ++;; it under the terms of the GNU General Public License as published by ++;; the Free Software Foundation; either version 2 of the License, or ++;; (at your option) any later version. ++;; ++;; This program is distributed in the hope that it will be useful, ++;; but WITHOUT ANY WARRANTY; without even the implied warranty of ++;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++;; GNU General Public License for more details. ++;; ++;; You should have received a copy of the GNU General Public License ++;; along with this program; if not, write to the Free Software ++;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. ++ ++ ++;; True if the operand is a memory reference which contains an ++;; Address consisting of a single pointer register ++(define_predicate "avr32_indirect_register_operand" ++ (and (match_code "mem") ++ (match_test "register_operand(XEXP(op, 0), SImode)"))) ++ ++ ++ ++;; Address expression with a base pointer offset with ++;; a register displacement ++(define_predicate "avr32_indexed_memory_operand" ++ (and (match_code "mem") ++ (match_test "GET_CODE(XEXP(op, 0)) == PLUS")) ++ { ++ ++ rtx op0 = XEXP(XEXP(op, 0), 0); ++ rtx op1 = XEXP(XEXP(op, 0), 1); ++ ++ return ((avr32_address_register_rtx_p (op0, 0) ++ && avr32_legitimate_index_p (GET_MODE(op), op1, 0)) ++ || (avr32_address_register_rtx_p (op1, 0) ++ && avr32_legitimate_index_p (GET_MODE(op), op0, 0))); ++ ++ }) ++ ++;; Operand suitable for the ld.sb instruction ++(define_predicate "load_sb_memory_operand" ++ (ior (match_operand 0 "avr32_indirect_register_operand") ++ (match_operand 0 "avr32_indexed_memory_operand"))) ++ ++ ++;; Operand suitable as operand to insns sign extending QI values ++(define_predicate "extendqi_operand" ++ (ior (match_operand 0 "load_sb_memory_operand") ++ (match_operand 0 "register_operand"))) ++ ++(define_predicate "post_inc_memory_operand" ++ (and (match_code "mem") ++ (match_test "(GET_CODE(XEXP(op, 0)) == POST_INC) ++ && REG_P(XEXP(XEXP(op, 0), 0))"))) ++ ++(define_predicate "pre_dec_memory_operand" ++ (and (match_code "mem") ++ (match_test "(GET_CODE(XEXP(op, 0)) == PRE_DEC) ++ && REG_P(XEXP(XEXP(op, 0), 0))"))) ++ ++;; Operand suitable for add instructions ++(define_predicate "avr32_add_operand" ++ (ior (match_operand 0 "register_operand") ++ (and (match_operand 0 "immediate_operand") ++ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'I', \"Is21\")")))) ++ ++;; Operand is a power of two immediate ++(define_predicate "power_of_two_operand" ++ (match_code "const_int") ++{ ++ HOST_WIDE_INT value = INTVAL (op); ++ ++ return value != 0 && (value & (value - 1)) == 0; ++}) ++ ++;; Operand is a multiple of 8 immediate ++(define_predicate "multiple_of_8_operand" ++ (match_code "const_int") ++{ ++ HOST_WIDE_INT value = INTVAL (op); ++ ++ return (value & 0x7) == 0 ; ++}) ++ ++;; Operand is a multiple of 16 immediate ++(define_predicate "multiple_of_16_operand" ++ (match_code "const_int") ++{ ++ HOST_WIDE_INT value = INTVAL (op); ++ ++ return (value & 0xf) == 0 ; ++}) ++ ++;; Operand is a mask used for masking away upper bits of a reg ++(define_predicate "avr32_mask_upper_bits_operand" ++ (match_code "const_int") ++{ ++ HOST_WIDE_INT value = INTVAL (op) + 1; ++ ++ return value != 1 && value != 0 && (value & (value - 1)) == 0; ++}) ++ ++ ++;; Operand suitable for mul instructions ++(define_predicate "avr32_mul_operand" ++ (ior (match_operand 0 "register_operand") ++ (and (match_operand 0 "immediate_operand") ++ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ks08\")")))) ++ ++;; True for logical binary operators. ++(define_predicate "logical_binary_operator" ++ (match_code "ior,xor,and")) ++ ++;; True for logical shift operators ++(define_predicate "logical_shift_operator" ++ (match_code "ashift,lshiftrt")) ++ ++;; True for shift operand for logical and, or and eor insns ++(define_predicate "avr32_logical_shift_operand" ++ (and (match_code "ashift,lshiftrt") ++ (ior (and (match_test "GET_CODE(XEXP(op, 1)) == CONST_INT") ++ (match_test "register_operand(XEXP(op, 0), GET_MODE(XEXP(op, 0)))")) ++ (and (match_test "GET_CODE(XEXP(op, 0)) == CONST_INT") ++ (match_test "register_operand(XEXP(op, 1), GET_MODE(XEXP(op, 1)))")))) ++ { ++ return 1; ++ } ++ ) ++ ++ ++;; Predicate for second operand to and, ior and xor insn patterns ++(define_predicate "avr32_logical_insn_operand" ++ (ior (match_operand 0 "register_operand") ++ (match_operand 0 "avr32_logical_shift_operand")) ++ { ++ return 1; ++ } ++) ++ ++ ++;; True for avr32 comparison operators ++(define_predicate "avr32_comparison_operator" ++ (ior (match_code "eq, ne, gt, ge, lt, le, gtu, geu, ltu, leu") ++ (and (match_code "unspec") ++ (match_test "(XINT(op, 1) == UNSPEC_COND_MI) ++ || (XINT(op, 1) == UNSPEC_COND_PL)")))) ++ ++;; True for avr32 comparison operand ++(define_predicate "avr32_comparison_operand" ++ (ior (and (match_code "eq, ne, gt, ge, lt, le, gtu, geu, ltu, leu") ++ (match_test "(rtx_equal_p (XEXP(op,0), cc0_rtx) && rtx_equal_p (XEXP(op,1), const0_rtx))")) ++ (and (match_code "unspec") ++ (match_test "(XINT(op, 1) == UNSPEC_COND_MI) ++ || (XINT(op, 1) == UNSPEC_COND_PL)")))) ++ ++;; True if this is a const_int with one bit set ++(define_predicate "one_bit_set_operand" ++ (match_code "const_int") ++ { ++ int i; ++ int value; ++ int ones = 0; ++ ++ value = INTVAL(op); ++ for ( i = 0 ; i < 32; i++ ){ ++ if ( value & ( 1 << i ) ){ ++ ones++; ++ } ++ } ++ ++ return ( ones == 1 ); ++ }) ++ ++ ++;; True if this is a const_int with one bit cleared ++(define_predicate "one_bit_cleared_operand" ++ (match_code "const_int") ++ { ++ int i; ++ int value; ++ int zeroes = 0; ++ ++ value = INTVAL(op); ++ for ( i = 0 ; i < 32; i++ ){ ++ if ( !(value & ( 1 << i )) ){ ++ zeroes++; ++ } ++ } ++ ++ return ( zeroes == 1 ); ++ }) ++ ++ ++;; True if this is a register or immediate operand ++(define_predicate "register_immediate_operand" ++ (ior (match_operand 0 "register_operand") ++ (match_operand 0 "immediate_operand"))) ++ ++ ++;; True is this is an operand containing a label_ref ++(define_predicate "avr32_label_ref_operand" ++ (and (match_code "mem") ++ (match_test "avr32_find_symbol(op) ++ && (GET_CODE(avr32_find_symbol(op)) == LABEL_REF)"))) ++ ++;; True is this is a valid symbol pointing to the constant pool ++(define_predicate "avr32_const_pool_operand" ++ (and (match_code "symbol_ref") ++ (match_test "CONSTANT_POOL_ADDRESS_P(op)")) ++ { ++ return (flag_pic ? (!(symbol_mentioned_p (get_pool_constant (op)) ++ || label_mentioned_p (get_pool_constant (op))) ++ || avr32_got_mentioned_p(get_pool_constant (op))) ++ : true); ++ } ++) ++ ++;; True is this is a memory reference to the constant or mini pool ++(define_predicate "avr32_const_pool_ref_operand" ++ (ior (match_operand 0 "avr32_label_ref_operand") ++ (and (match_code "mem") ++ (match_test "avr32_const_pool_operand(XEXP(op,0), GET_MODE(XEXP(op,0)))")))) ++ ++ ++;; Legal source operand for movti insns ++(define_predicate "avr32_movti_src_operand" ++ (ior (match_operand 0 "avr32_const_pool_ref_operand") ++ (ior (ior (match_operand 0 "register_immediate_operand") ++ (match_operand 0 "avr32_indirect_register_operand")) ++ (match_operand 0 "post_inc_memory_operand")))) ++ ++;; Legal destination operand for movti insns ++(define_predicate "avr32_movti_dst_operand" ++ (ior (ior (match_operand 0 "register_operand") ++ (match_operand 0 "avr32_indirect_register_operand")) ++ (match_operand 0 "pre_dec_memory_operand"))) ++ ++ ++;; True is this is a k12 offseted memory operand ++(define_predicate "avr32_k12_memory_operand" ++ (and (match_code "mem") ++ (ior (match_test "REG_P(XEXP(op, 0))") ++ (match_test "GET_CODE(XEXP(op, 0)) == PLUS ++ && REG_P(XEXP(XEXP(op, 0), 0)) ++ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT) ++ && (CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(XEXP(op, 0), 0)), ++ 'K', (mode == SImode) ? \"Ks14\" : ((mode == HImode) ? \"Ks13\" : \"Ks12\")))")))) ++ ++;; True is this is a memory operand with an immediate displacement ++(define_predicate "avr32_imm_disp_memory_operand" ++ (and (match_code "mem") ++ (match_test "GET_CODE(XEXP(op, 0)) == PLUS ++ && REG_P(XEXP(XEXP(op, 0), 0)) ++ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT)"))) ++ ++;; True is this is a bswap operand ++(define_predicate "avr32_bswap_operand" ++ (ior (match_operand 0 "avr32_k12_memory_operand") ++ (match_operand 0 "register_operand"))) ++ ++;; True is this is a valid coprocessor insn memory operand ++(define_predicate "avr32_cop_memory_operand" ++ (and (match_operand 0 "memory_operand") ++ (not (match_test "GET_CODE(XEXP(op, 0)) == PLUS ++ && REG_P(XEXP(XEXP(op, 0), 0)) ++ && (GET_CODE(XEXP(XEXP(op, 0), 1)) == CONST_INT) ++ && !(CONST_OK_FOR_CONSTRAINT_P(INTVAL(XEXP(XEXP(op, 0), 0)), 'K', \"Ku10\"))")))) ++ ++;; True is this is a valid source/destination operand ++;; for moving values to/from a coprocessor ++(define_predicate "avr32_cop_move_operand" ++ (ior (match_operand 0 "register_operand") ++ (match_operand 0 "avr32_cop_memory_operand"))) ++ ++ ++;; True is this is a valid extract byte offset for use in ++;; load extracted index insns ++(define_predicate "avr32_extract_shift_operand" ++ (and (match_operand 0 "const_int_operand") ++ (match_test "(INTVAL(op) == 0) || (INTVAL(op) == 8) ++ || (INTVAL(op) == 16) || (INTVAL(op) == 24)"))) ++ ++;; True is this is a floating-point register ++(define_predicate "avr32_fp_register_operand" ++ (and (match_operand 0 "register_operand") ++ (match_test "REGNO_REG_CLASS(REGNO(op)) == FP_REGS"))) ++ ++;; True is this is valid avr32 symbol operand ++(define_predicate "avr32_symbol_operand" ++ (ior (match_code "label_ref, symbol_ref") ++ (and (match_code "const") ++ (match_test "avr32_find_symbol(op)")))) ++ ++;; True is this is valid operand for the lda.w and call pseudo insns ++(define_predicate "avr32_address_operand" ++ (and (match_code "label_ref, symbol_ref") ++ (ior (match_test "TARGET_HAS_ASM_ADDR_PSEUDOS") ++ (match_test "flag_pic")) )) ++ ++;; True if this is a avr32 call operand ++(define_predicate "avr32_call_operand" ++ (ior (ior (match_operand 0 "register_operand") ++ (ior (match_operand 0 "avr32_const_pool_ref_operand") ++ (match_operand 0 "avr32_address_operand"))) ++ (match_test "SYMBOL_REF_RCALL_FUNCTION_P(op)"))) ++ ++;; Return true for operators performing ALU operations ++ ++(define_predicate "alu_operator" ++ (match_code "ior, xor, and, plus, minus, ashift, lshiftrt, ashiftrt")) ++ ++(define_predicate "avr32_add_shift_immediate_operand" ++ (and (match_operand 0 "immediate_operand") ++ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ku02\")"))) ++ ++(define_predicate "avr32_cond_register_immediate_operand" ++ (ior (match_operand 0 "register_operand") ++ (and (match_operand 0 "immediate_operand") ++ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'K', \"Ks08\")")))) ++ ++(define_predicate "avr32_cond_immediate_operand" ++ (and (match_operand 0 "immediate_operand") ++ (match_test "CONST_OK_FOR_CONSTRAINT_P(INTVAL(op), 'I', \"Is08\")"))) +diff -Nrup gcc-4.2.1/gcc/config/avr32/simd.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/simd.md +--- gcc-4.2.1/gcc/config/avr32/simd.md 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/simd.md 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,145 @@ ++;; AVR32 machine description file for SIMD instructions. ++;; Copyright 2003-2006 Atmel Corporation. ++;; ++;; Written by Ronny Pedersen, Atmel Norway, ++;; ++;; This file is part of GCC. ++;; ++;; This program is free software; you can redistribute it and/or modify ++;; it under the terms of the GNU General Public License as published by ++;; the Free Software Foundation; either version 2 of the License, or ++;; (at your option) any later version. ++;; ++;; This program is distributed in the hope that it will be useful, ++;; but WITHOUT ANY WARRANTY; without even the implied warranty of ++;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++;; GNU General Public License for more details. ++;; ++;; You should have received a copy of the GNU General Public License ++;; along with this program; if not, write to the Free Software ++;; Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. ++ ++;; -*- Mode: Scheme -*- ++ ++ ++;; Vector modes ++(define_mode_macro VECM [V2HI V4QI]) ++(define_mode_attr size [(V2HI "h") (V4QI "b")]) ++ ++(define_insn "add3" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (plus:VECM (match_operand:VECM 1 "register_operand" "r") ++ (match_operand:VECM 2 "register_operand" "r")))] ++ "TARGET_SIMD" ++ "padd.\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++ ++(define_insn "sub3" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (minus:VECM (match_operand:VECM 1 "register_operand" "r") ++ (match_operand:VECM 2 "register_operand" "r")))] ++ "TARGET_SIMD" ++ "psub.\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++ ++(define_insn "abs2" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (abs:VECM (match_operand:VECM 1 "register_operand" "r")))] ++ "TARGET_SIMD" ++ "pabs.s\t%0, %1" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "ashl3" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (ashift:VECM (match_operand:VECM 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku04")))] ++ "TARGET_SIMD" ++ "plsl.\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "ashr3" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (ashiftrt:VECM (match_operand:VECM 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku04")))] ++ "TARGET_SIMD" ++ "pasr.\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "lshr3" ++ [(set (match_operand:VECM 0 "register_operand" "=r") ++ (lshiftrt:VECM (match_operand:VECM 1 "register_operand" "r") ++ (match_operand:SI 2 "immediate_operand" "Ku04")))] ++ "TARGET_SIMD" ++ "plsr.\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "smaxv2hi3" ++ [(set (match_operand:V2HI 0 "register_operand" "=r") ++ (smax:V2HI (match_operand:V2HI 1 "register_operand" "r") ++ (match_operand:V2HI 2 "register_operand" "r")))] ++ ++ "TARGET_SIMD" ++ "pmax.sh\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "sminv2hi3" ++ [(set (match_operand:V2HI 0 "register_operand" "=r") ++ (smin:V2HI (match_operand:V2HI 1 "register_operand" "r") ++ (match_operand:V2HI 2 "register_operand" "r")))] ++ ++ "TARGET_SIMD" ++ "pmin.sh\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "umaxv4qi3" ++ [(set (match_operand:V4QI 0 "register_operand" "=r") ++ (umax:V4QI (match_operand:V4QI 1 "register_operand" "r") ++ (match_operand:V4QI 2 "register_operand" "r")))] ++ ++ "TARGET_SIMD" ++ "pmax.ub\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "uminv4qi3" ++ [(set (match_operand:V4QI 0 "register_operand" "=r") ++ (umin:V4QI (match_operand:V4QI 1 "register_operand" "r") ++ (match_operand:V4QI 2 "register_operand" "r")))] ++ ++ "TARGET_SIMD" ++ "pmin.ub\t%0, %1, %2" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++ ++(define_insn "addsubv2hi" ++ [(set (match_operand:V2HI 0 "register_operand" "=r") ++ (vec_concat:V2HI ++ (plus:HI (match_operand:HI 1 "register_operand" "r") ++ (match_operand:HI 2 "register_operand" "r")) ++ (minus:HI (match_dup 1) (match_dup 2))))] ++ "TARGET_SIMD" ++ "paddsub.h\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) ++ ++(define_insn "subaddv2hi" ++ [(set (match_operand:V2HI 0 "register_operand" "=r") ++ (vec_concat:V2HI ++ (minus:HI (match_operand:HI 1 "register_operand" "r") ++ (match_operand:HI 2 "register_operand" "r")) ++ (plus:HI (match_dup 1) (match_dup 2))))] ++ "TARGET_SIMD" ++ "psubadd.h\t%0, %1:b, %2:b" ++ [(set_attr "length" "4") ++ (set_attr "type" "alu")]) +diff -Nrup gcc-4.2.1/gcc/config/avr32/sync.md gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/sync.md +--- gcc-4.2.1/gcc/config/avr32/sync.md 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/sync.md 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,175 @@ ++;;================================================================= ++;; Atomic operations ++;;================================================================= ++ ++ ++(define_insn "sync_compare_and_swapsi" ++ [(set (match_operand:SI 0 "register_operand" "=&r,&r") ++ (match_operand:SI 1 "memory_operand" "+RKs16,+RKs16")) ++ (set (match_dup 1) ++ (unspec_volatile:SI ++ [(match_dup 1) ++ (match_operand:SI 2 "register_immediate_operand" "r,Ks21") ++ (match_operand:SI 3 "register_operand" "r,r")] ++ VUNSPEC_SYNC_CMPXCHG)) ] ++ "" ++ "0: ++ ssrf\t5 ++ ld.w\t%0,%1 ++ cp.w\t%0,%2 ++ brne\t0f ++ stcond\t%1, %3 ++ brne\t0b ++ 0: ++ " ++ [(set_attr "length" "16,18") ++ (set_attr "cc" "clobber")] ++ ) ++ ++ ++(define_code_macro atomic_op [plus minus and ior xor]) ++(define_code_attr atomic_asm_insn [(plus "add") (minus "sub") (and "and") (ior "or") (xor "eor")]) ++(define_code_attr atomic_insn [(plus "add") (minus "sub") (and "and") (ior "ior") (xor "xor")]) ++ ++(define_insn "sync_loadsi" ++ [(set (match_operand:SI 0 "register_operand" "=r") ++ (unspec_volatile:SI ++ [(match_operand:SI 1 "memory_operand" "RKs16") ++ (label_ref (match_operand 2 "" ""))] ++ VUNSPEC_SYNC_SET_LOCK_AND_LOAD) )] ++ "" ++ "%2: ++ ssrf\t5 ++ ld.w\t%0,%1" ++ [(set_attr "length" "6") ++ (set_attr "cc" "clobber")] ++ ) ++ ++(define_insn "sync_store_if_lock" ++ [(set (match_operand:SI 0 "memory_operand" "=RKs16") ++ (unspec_volatile:SI ++ [(match_operand:SI 1 "register_operand" "r") ++ (label_ref (match_operand 2 "" ""))] ++ VUNSPEC_SYNC_STORE_IF_LOCK) )] ++ "" ++ "stcond\t%0, %1 ++ brne\t%2" ++ [(set_attr "length" "6") ++ (set_attr "cc" "clobber")] ++ ) ++ ++ ++(define_expand "sync_si" ++ [(set (match_dup 2) ++ (unspec_volatile:SI ++ [(match_operand:SI 0 "memory_operand" "") ++ (match_dup 3)] ++ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) ++ (set (match_dup 2) ++ (atomic_op:SI (match_dup 2) ++ (match_operand:SI 1 "register_immediate_operand" ""))) ++ (set (match_dup 0) ++ (unspec_volatile:SI ++ [(match_dup 2) ++ (match_dup 3)] ++ VUNSPEC_SYNC_STORE_IF_LOCK) )] ++ "" ++ { ++ operands[2] = gen_reg_rtx (SImode); ++ operands[3] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); ++ } ++ ) ++ ++ ++ ++(define_expand "sync_old_si" ++ [(set (match_operand:SI 0 "register_operand" "") ++ (unspec_volatile:SI ++ [(match_operand:SI 1 "memory_operand" "") ++ (match_dup 4)] ++ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) ++ (set (match_dup 3) ++ (atomic_op:SI (match_dup 0) ++ (match_operand:SI 2 "register_immediate_operand" ""))) ++ (set (match_dup 1) ++ (unspec_volatile:SI ++ [(match_dup 3) ++ (match_dup 4)] ++ VUNSPEC_SYNC_STORE_IF_LOCK) )] ++ "" ++ { ++ operands[3] = gen_reg_rtx (SImode); ++ operands[4] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); ++ } ++ ) ++ ++(define_expand "sync_new_si" ++ [(set (match_operand:SI 0 "register_operand" "") ++ (unspec_volatile:SI ++ [(match_operand:SI 1 "memory_operand" "") ++ (match_dup 3)] ++ VUNSPEC_SYNC_SET_LOCK_AND_LOAD)) ++ (set (match_dup 0) ++ (atomic_op:SI (match_dup 0) ++ (match_operand:SI 2 "register_immediate_operand" ""))) ++ (set (match_dup 1) ++ (unspec_volatile:SI ++ [(match_dup 0) ++ (match_dup 3)] ++ VUNSPEC_SYNC_STORE_IF_LOCK) )] ++ "" ++ { ++ operands[3] = gen_rtx_LABEL_REF(Pmode, gen_label_rtx ()); ++ } ++ ) ++ ++ ++;(define_insn "sync_si" ++; [(set (match_operand:SI 0 "memory_operand" "+RKs16") ++; (unspec_volatile:SI ++; [(atomic_op:SI (match_dup 0) ++; (match_operand:SI 1 "register_operand" "r"))] ++; VUNSPEC_SYNC_CMPXCHG)) ++; (clobber (match_scratch:SI 2 "=&r"))] ++; "" ++; "0: ++; ssrf\t5 ++; ld.w\t%2,%0 ++; \t%2,%1 ++; stcond\t%0, %2 ++; brne\t0b ++; " ++; [(set_attr "length" "14") ++; (set_attr "cc" "clobber")] ++; ) ++; ++;(define_insn "sync_new_si" ++; [(set (match_operand:SI 1 "memory_operand" "+RKs16") ++; (unspec_volatile:SI ++; [(atomic_op:SI (match_dup 1) ++; (match_operand:SI 2 "register_operand" "r"))] ++; VUNSPEC_SYNC_CMPXCHG)) ++; (set (match_operand:SI 0 "register_operand" "=&r") ++; (atomic_op:SI (match_dup 1) ++; (match_dup 2)))] ++; "" ++; "0: ++; ssrf\t5 ++; ld.w\t%0,%1 ++; \t%0,%2 ++; stcond\t%1, %0 ++; brne\t0b ++; " ++; [(set_attr "length" "14") ++; (set_attr "cc" "clobber")] ++; ) ++ ++(define_insn "sync_lock_test_and_setsi" ++ [ (set (match_operand:SI 0 "register_operand" "=&r") ++ (match_operand:SI 1 "memory_operand" "+RKu00")) ++ (set (match_dup 1) ++ (match_operand:SI 2 "register_operand" "r")) ] ++ "" ++ "xchg\t%0, %p1, %2" ++ [(set_attr "length" "4")] ++ ) +diff -Nrup gcc-4.2.1/gcc/config/avr32/t-avr32 gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-avr32 +--- gcc-4.2.1/gcc/config/avr32/t-avr32 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-avr32 2007-09-28 10:33:00.000000000 +0200 +@@ -0,0 +1,76 @@ ++ ++MD_INCLUDES= $(srcdir)/config/avr32/avr32.md \ ++ $(srcdir)/config/avr32/sync.md \ ++ $(srcdir)/config/avr32/fpcp.md \ ++ $(srcdir)/config/avr32/simd.md \ ++ $(srcdir)/config/avr32/predicates.md ++ ++s-config s-conditions s-flags s-codes s-constants s-emit s-recog s-preds \ ++ s-opinit s-extract s-peep s-attr s-attrtab s-output: $(MD_INCLUDES) ++ ++# We want fine grained libraries, so use the new code ++# to build the floating point emulation libraries. ++FPBIT = fp-bit.c ++DPBIT = dp-bit.c ++ ++LIB1ASMSRC = avr32/lib1funcs.S ++LIB1ASMFUNCS = _avr32_f64_mul _avr32_f64_addsub _avr32_f64_addsub_fast _avr32_f64_to_u32 \ ++ _avr32_f64_to_s32 _avr32_f64_to_u64 _avr32_f64_to_s64 _avr32_u32_to_f64 \ ++ _avr32_s32_to_f64 _avr32_f64_cmp_eq _avr32_f64_cmp_ge _avr32_f64_cmp_lt \ ++ _avr32_f32_cmp_eq _avr32_f32_cmp_ge _avr32_f32_cmp_lt _avr32_f64_div \ ++ _avr32_f32_div _avr32_f32_div_fast _avr32_f32_addsub _avr32_f32_addsub_fast \ ++ _avr32_f32_mul _avr32_s32_to_f32 _avr32_u32_to_f32 _avr32_f32_to_s32 \ ++ _avr32_f32_to_u32 _avr32_f32_to_f64 _avr32_f64_to_f32 ++ ++#LIB2FUNCS_EXTRA += $(srcdir)/config/avr32/lib2funcs.S ++ ++MULTILIB_OPTIONS = march=ap/march=uc ++MULTILIB_DIRNAMES = ap uc ++MULTILIB_EXCEPTIONS = ++MULTILIB_MATCHES = march?ap=mcpu?ap7000 ++MULTILIB_MATCHES += march?ap=mcpu?ap7010 ++MULTILIB_MATCHES += march?ap=mcpu?ap7020 ++MULTILIB_MATCHES += march?uc=mcpu?uc3a0256 ++MULTILIB_MATCHES += march?uc=mcpu?uc3a0512 ++MULTILIB_MATCHES += march?uc=mcpu?uc3a1128 ++MULTILIB_MATCHES += march?uc=mcpu?uc3a1256 ++MULTILIB_MATCHES += march?uc=mcpu?uc3a1512 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b064 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b0128 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b0256 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b164 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b1128 ++MULTILIB_MATCHES += march?uc=mcpu?uc3b1256 ++MULTILIB_MATCHES += march?ap=mpart?ap7000 ++MULTILIB_MATCHES += march?ap=mpart?ap7010 ++MULTILIB_MATCHES += march?ap=mpart?ap7020 ++MULTILIB_MATCHES += march?uc=mpart?uc3a0256 ++MULTILIB_MATCHES += march?uc=mpart?uc3a0512 ++MULTILIB_MATCHES += march?uc=mpart?uc3a1128 ++MULTILIB_MATCHES += march?uc=mpart?uc3a1256 ++MULTILIB_MATCHES += march?uc=mpart?uc3a1512 ++MULTILIB_MATCHES += march?uc=mpart?uc3b064 ++MULTILIB_MATCHES += march?uc=mpart?uc3b0128 ++MULTILIB_MATCHES += march?uc=mpart?uc3b0256 ++MULTILIB_MATCHES += march?uc=mpart?uc3b164 ++MULTILIB_MATCHES += march?uc=mpart?uc3b1128 ++MULTILIB_MATCHES += march?uc=mpart?uc3b1256 ++ ++EXTRA_MULTILIB_PARTS = crtbegin.o crtbeginS.o crtend.o crtendS.o crti.o crtn.o ++ ++CRTSTUFF_T_CFLAGS = -mrelax ++CRTSTUFF_T_CFLAGS_S = -mrelax -fPIC ++TARGET_LIBGCC2_CFLAGS += -mrelax ++ ++LIBGCC = stmp-multilib ++INSTALL_LIBGCC = install-multilib ++ ++fp-bit.c: $(srcdir)/config/fp-bit.c ++ echo '#define FLOAT' > fp-bit.c ++ cat $(srcdir)/config/fp-bit.c >> fp-bit.c ++ ++dp-bit.c: $(srcdir)/config/fp-bit.c ++ cat $(srcdir)/config/fp-bit.c > dp-bit.c ++ ++ ++ +diff -Nrup gcc-4.2.1/gcc/config/avr32/t-elf gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-elf +--- gcc-4.2.1/gcc/config/avr32/t-elf 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/t-elf 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,16 @@ ++ ++# Assemble startup files. ++$(T)crti.o: $(srcdir)/config/avr32/crti.asm $(GCC_PASSES) ++ $(GCC_FOR_TARGET) $(CRTSTUFF_CFLAGS) $(CRTSTUFF_T_CFLAGS) $(INCLUDES) \ ++ -c -o $(T)crti.o -x assembler-with-cpp $(srcdir)/config/avr32/crti.asm ++ ++$(T)crtn.o: $(srcdir)/config/avr32/crtn.asm $(GCC_PASSES) ++ $(GCC_FOR_TARGET) $(CRTSTUFF_CFLAGS) $(CRTSTUFF_T_CFLAGS) $(INCLUDES) \ ++ -c -o $(T)crtn.o -x assembler-with-cpp $(srcdir)/config/avr32/crtn.asm ++ ++ ++# Build the libraries for both hard and soft floating point ++EXTRA_MULTILIB_PARTS = crtbegin.o crtbeginS.o crtend.o crtendS.o crti.o crtn.o ++ ++LIBGCC = stmp-multilib ++INSTALL_LIBGCC = install-multilib +diff -Nrup gcc-4.2.1/gcc/config/avr32/uclinux-elf.h gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/uclinux-elf.h +--- gcc-4.2.1/gcc/config/avr32/uclinux-elf.h 1970-01-01 01:00:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/avr32/uclinux-elf.h 2007-05-07 14:29:10.000000000 +0200 +@@ -0,0 +1,20 @@ ++ ++/* Run-time Target Specification. */ ++#undef TARGET_VERSION ++#define TARGET_VERSION fputs (" (AVR32 uClinux with ELF)", stderr) ++ ++/* We don't want a .jcr section on uClinux. As if this makes a difference... */ ++#define TARGET_USE_JCR_SECTION 0 ++ ++/* Here we go. Drop the crtbegin/crtend stuff completely. */ ++#undef STARTFILE_SPEC ++#define STARTFILE_SPEC \ ++ "%{!shared: %{pg:gcrt1.o%s} %{!pg:%{p:gcrt1.o%s}" \ ++ " %{!p:%{profile:gcrt1.o%s}" \ ++ " %{!profile:crt1.o%s}}}} crti.o%s" ++ ++#undef ENDFILE_SPEC ++#define ENDFILE_SPEC "crtn.o%s" ++ ++#undef TARGET_DEFAULT ++#define TARGET_DEFAULT (AVR32_FLAG_NO_INIT_GOT) +diff -Nrup gcc-4.2.1/gcc/config/host-linux.c gcc-4.2.1.atmel.1.3.2/gcc/config/host-linux.c +--- gcc-4.2.1/gcc/config/host-linux.c 2005-08-01 19:43:33.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config/host-linux.c 2007-05-07 14:29:15.000000000 +0200 +@@ -26,6 +26,9 @@ + #include "hosthooks.h" + #include "hosthooks-def.h" + ++#ifndef SSIZE_MAX ++#define SSIZE_MAX LONG_MAX ++#endif + + /* Linux has a feature called exec-shield-randomize that perturbs the + address of non-fixed mapped segments by a (relatively) small amount. +diff -Nrup gcc-4.2.1/gcc/config.gcc gcc-4.2.1.atmel.1.3.2/gcc/config.gcc +--- gcc-4.2.1/gcc/config.gcc 2007-02-03 06:25:20.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/config.gcc 2007-09-28 10:33:09.000000000 +0200 +@@ -782,6 +782,24 @@ avr-*-*) + tm_file="avr/avr.h dbxelf.h" + use_fixproto=yes + ;; ++avr32*-*-linux*) ++ tm_file="dbxelf.h elfos.h linux.h avr32/linux-elf.h avr32/avr32.h " ++ tmake_file="t-linux avr32/t-avr32 avr32/t-elf" ++ extra_parts="crtbegin.o crtbeginS.o crtend.o crtendS.o" ++ extra_modes=avr32/avr32-modes.def ++ gnu_ld=yes ++ ;; ++avr32*-*-uclinux*) ++ tm_file="dbxelf.h elfos.h linux.h avr32/linux-elf.h avr32/uclinux-elf.h avr32/avr32.h" ++ tmake_file="t-linux avr32/t-avr32 avr32/t-elf" ++ extra_modes=avr32/avr32-modes.def ++ gnu_ld=yes ++ ;; ++avr32-*-*) ++ tm_file="dbxelf.h elfos.h avr32/avr32.h avr32/avr32-elf.h" ++ tmake_file="avr32/t-avr32 avr32/t-elf" ++ extra_modes=avr32/avr32-modes.def ++ ;; + bfin*-elf*) + tm_file="${tm_file} dbxelf.h elfos.h bfin/elf.h" + tmake_file=bfin/t-bfin-elf +@@ -1682,6 +1700,9 @@ pdp11-*-bsd) + pdp11-*-*) + use_fixproto=yes + ;; ++avr-*-*) ++ use_fixproto=yes ++ ;; + # port not yet contributed + #powerpc-*-openbsd*) + # tmake_file="${tmake_file} rs6000/t-fprules rs6000/t-fprules-fpbit " +@@ -2718,6 +2739,32 @@ case "${target}" in + fi + ;; + ++ avr32*-*-*) ++ supported_defaults="part arch" ++ ++ case "$with_part" in ++ "" \ ++ | "ap7000" | "ap7010" | "ap7020" | "uc3a0256" | "uc3a0512" | "uc3a1128" | "uc3a1256" | "uc3a1512" ) ++ # OK ++ ;; ++ *) ++ echo "Unknown part used in --with-part=$with_part" 1>&2 ++ exit 1 ++ ;; ++ esac ++ ++ case "$with_arch" in ++ "" \ ++ | "ap" | "uc") ++ # OK ++ ;; ++ *) ++ echo "Unknown arch used in --with-arch=$with_arch" 1>&2 ++ exit 1 ++ ;; ++ esac ++ ;; ++ + fr*-*-*linux*) + supported_defaults=cpu + case "$with_cpu" in +diff -Nrup gcc-4.2.1/gcc/doc/extend.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/extend.texi +--- gcc-4.2.1/gcc/doc/extend.texi 2007-03-12 23:10:12.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/doc/extend.texi 2007-09-28 10:32:30.000000000 +0200 +@@ -1981,7 +1981,7 @@ this attribute to work correctly. + + @item interrupt + @cindex interrupt handler functions +-Use this attribute on the ARM, AVR, C4x, CRX, M32C, M32R/D, MS1, and Xstormy16 ++Use this attribute on the ARM, AVR, AVR32, C4x, CRX, M32C, M32R/D, MS1, and Xstormy16 + ports to indicate that the specified function is an interrupt handler. + The compiler will generate function entry and exit sequences suitable + for use in an interrupt handler when this attribute is present. +@@ -2000,6 +2000,15 @@ void f () __attribute__ ((interrupt ("IR + + Permissible values for this parameter are: IRQ, FIQ, SWI, ABORT and UNDEF@. + ++Note, for the AVR32, you can specify which banking scheme is used for ++the interrupt mode this interrupt handler is used in like this: ++ ++@smallexample ++void f () __attribute__ ((interrupt ("FULL"))); ++@end smallexample ++ ++Permissible values for this parameter are: FULL, HALF, NONE and UNDEF. ++ + @item interrupt_handler + @cindex interrupt handler functions on the Blackfin, m68k, H8/300 and SH processors + Use this attribute on the Blackfin, m68k, H8/300, H8/300H, H8S, and SH to +@@ -6167,6 +6176,7 @@ instructions, but allow the compiler to + @menu + * Alpha Built-in Functions:: + * ARM Built-in Functions:: ++* AVR32 Built-in Functions:: + * Blackfin Built-in Functions:: + * FR-V Built-in Functions:: + * X86 Built-in Functions:: +@@ -6405,6 +6415,54 @@ long long __builtin_arm_wxor (long long, + long long __builtin_arm_wzero () + @end smallexample + ++@node AVR32 Built-in Functions ++@subsection AVR32 Built-in Functions ++ ++ ++@smallexample ++ ++int __builtin_sats (int /*Rd*/,int /*sa*/, int /*bn*/) ++int __builtin_satu (int /*Rd*/,int /*sa*/, int /*bn*/) ++int __builtin_satrnds (int /*Rd*/,int /*sa*/, int /*bn*/) ++int __builtin_satrndu (int /*Rd*/,int /*sa*/, int /*bn*/) ++short __builtin_mulsathh_h (short, short) ++int __builtin_mulsathh_w (short, short) ++short __builtin_mulsatrndhh_h (short, short) ++int __builtin_mulsatrndwh_w (int, short) ++int __builtin_mulsatwh_w (int, short) ++int __builtin_macsathh_w (int, short, short) ++short __builtin_satadd_h (short, short) ++short __builtin_satsub_h (short, short) ++int __builtin_satadd_w (int, int) ++int __builtin_satsub_w (int, int) ++long long __builtin_mulwh_d(int, short) ++long long __builtin_mulnwh_d(int, short) ++long long __builtin_macwh_d(long long, int, short) ++long long __builtin_machh_d(long long, short, short) ++ ++void __builtin_musfr(int); ++int __builtin_mustr(void); ++int __builtin_mfsr(int /*Status Register Address*/) ++void __builtin_mtsr(int /*Status Register Address*/, int /*Value*/) ++int __builtin_mfdr(int /*Debug Register Address*/) ++void __builtin_mtdr(int /*Debug Register Address*/, int /*Value*/) ++void __builtin_cache(void * /*Address*/, int /*Cache Operation*/) ++void __builtin_sync(int /*Sync Operation*/) ++void __builtin_tlbr(void) ++void __builtin_tlbs(void) ++void __builtin_tlbw(void) ++void __builtin_breakpoint(void) ++int __builtin_xchg(void * /*Address*/, int /*Value*/ ) ++short __builtin_bswap_16(short) ++int __builtin_bswap_32(int) ++void __builtin_cop(int/*cpnr*/, int/*crd*/, int/*crx*/, int/*cry*/, int/*op*/) ++int __builtin_mvcr_w(int/*cpnr*/, int/*crs*/) ++void __builtin_mvrc_w(int/*cpnr*/, int/*crd*/, int/*value*/) ++long long __builtin_mvcr_d(int/*cpnr*/, int/*crs*/) ++void __builtin_mvrc_d(int/*cpnr*/, int/*crd*/, long long/*value*/) ++ ++@end smallexample ++ + @node Blackfin Built-in Functions + @subsection Blackfin Built-in Functions + +diff -Nrup gcc-4.2.1/gcc/doc/invoke.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/invoke.texi +--- gcc-4.2.1/gcc/doc/invoke.texi 2007-04-24 23:54:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/doc/invoke.texi 2007-09-28 10:32:30.000000000 +0200 +@@ -190,7 +190,7 @@ in the following sections. + -fno-default-inline -fvisibility-inlines-hidden @gol + -Wabi -Wctor-dtor-privacy @gol + -Wnon-virtual-dtor -Wreorder @gol +--Weffc++ -Wno-deprecated -Wstrict-null-sentinel @gol ++-Weffc++ -Wno-deprecated @gol + -Wno-non-template-friend -Wold-style-cast @gol + -Woverloaded-virtual -Wno-pmf-conversions @gol + -Wsign-promo} +@@ -588,6 +588,12 @@ Objective-C and Objective-C++ Dialects}. + -mauto-incdec -minmax -mlong-calls -mshort @gol + -msoft-reg-count=@var{count}} + ++@emph{AVR32 Options} ++@gccoptlist{-muse-rodata-section -mhard-float -msoft-float -mrelax @gol ++-mforce-double-align -mno-init-got -mpart=@var{part} -mcpu=@var{cpu} @gol ++-march=@var{arch} -mfast-float -masm-addr-pseudos -mno-asm-addr-pseudos -mno-pic ++} ++ + @emph{MCore Options} + @gccoptlist{-mhardlit -mno-hardlit -mdiv -mno-div -mrelax-immediates @gol + -mno-relax-immediates -mwide-bitfields -mno-wide-bitfields @gol +@@ -1868,14 +1874,6 @@ to filter out those warnings. + @opindex Wno-deprecated + Do not warn about usage of deprecated features. @xref{Deprecated Features}. + +-@item -Wstrict-null-sentinel @r{(C++ only)} +-@opindex Wstrict-null-sentinel +-Warn also about the use of an uncasted @code{NULL} as sentinel. When +-compiling only with GCC this is a valid sentinel, as @code{NULL} is defined +-to @code{__null}. Although it is a null pointer constant not a null pointer, +-it is guaranteed to of the same size as a pointer. But this use is +-not portable across different compilers. +- + @item -Wno-non-template-friend @r{(C++ only)} + @opindex Wno-non-template-friend + Disable warnings when non-templatized friend functions are declared +@@ -2732,13 +2730,11 @@ requiring @option{-O}. + If you want to warn about code which uses the uninitialized value of the + variable in its own initializer, use the @option{-Winit-self} option. + +-These warnings occur for individual uninitialized or clobbered +-elements of structure, union or array variables as well as for +-variables which are uninitialized or clobbered as a whole. They do +-not occur for variables or elements declared @code{volatile}. Because +-these warnings depend on optimization, the exact variables or elements +-for which there are warnings will depend on the precise optimization +-options and version of GCC used. ++These warnings occur only for variables that are candidates for ++register allocation. Therefore, they do not occur for a variable that ++is declared @code{volatile}, or whose address is taken, or whose size ++is other than 1, 2, 4 or 8 bytes. Also, they do not occur for ++structures, unions or arrays, even when they are in registers. + + Note that there may be no warning about a variable that is used only + to compute a value that itself is never used, because such +@@ -6201,10 +6197,6 @@ If number of candidates in the set is sm + we always try to remove unnecessary ivs from the set during its + optimization when a new iv is added to the set. + +-@item scev-max-expr-size +-Bound on size of expressions used in the scalar evolutions analyzer. +-Large expressions slow the analyzer. +- + @item vect-max-version-checks + The maximum number of runtime checks that can be performed when doing + loop versioning in the vectorizer. See option ftree-vect-loop-version +@@ -7402,7 +7394,7 @@ platform. + * ARC Options:: + * ARM Options:: + * AVR Options:: +-* Blackfin Options:: ++* AVR32 Options:: + * CRIS Options:: + * CRX Options:: + * Darwin Options:: +@@ -7867,81 +7859,68 @@ comply to the C standards, but it will p + size. + @end table + +-@node Blackfin Options +-@subsection Blackfin Options +-@cindex Blackfin Options ++@node AVR32 Options ++@subsection AVR32 Options ++@cindex AVR32 Options ++ ++These options are defined for AVR32 implementations: + + @table @gcctabopt +-@item -momit-leaf-frame-pointer +-@opindex momit-leaf-frame-pointer +-Don't keep the frame pointer in a register for leaf functions. This +-avoids the instructions to save, set up and restore frame pointers and +-makes an extra register available in leaf functions. The option +-@option{-fomit-frame-pointer} removes the frame pointer for all functions +-which might make debugging harder. ++@item -muse-rodata-section ++@opindex muse-rodata-section ++Use section @samp{.rodata} for read-only data instead of @samp{.text}. + +-@item -mspecld-anomaly +-@opindex mspecld-anomaly +-When enabled, the compiler will ensure that the generated code does not +-contain speculative loads after jump instructions. This option is enabled +-by default. +- +-@item -mno-specld-anomaly +-@opindex mno-specld-anomaly +-Don't generate extra code to prevent speculative loads from occurring. +- +-@item -mcsync-anomaly +-@opindex mcsync-anomaly +-When enabled, the compiler will ensure that the generated code does not +-contain CSYNC or SSYNC instructions too soon after conditional branches. +-This option is enabled by default. +- +-@item -mno-csync-anomaly +-@opindex mno-csync-anomaly +-Don't generate extra code to prevent CSYNC or SSYNC instructions from +-occurring too soon after a conditional branch. +- +-@item -mlow-64k +-@opindex mlow-64k +-When enabled, the compiler is free to take advantage of the knowledge that +-the entire program fits into the low 64k of memory. +- +-@item -mno-low-64k +-@opindex mno-low-64k +-Assume that the program is arbitrarily large. This is the default. ++@item -mrelax ++@opindex mrelax ++Enable relaxing in linker. This means that when the address of symbols ++are known at link time, the linker can optimize @samp{icall} and @samp{mcall} ++instructions into a @samp{rcall} instruction if possible. Loading the address ++of a symbol can also be optimized. ++ ++@item -mforce-double-align ++@opindex mforce-double-align ++Force double-word alignment for double-word memory accesses. ++ ++@item -mno-init-got ++@opindex mno-init-got ++Do not initialize the GOT register before using it when compiling PIC ++code. + +-@item -mid-shared-library +-@opindex mid-shared-library +-Generate code that supports shared libraries via the library ID method. +-This allows for execute in place and shared libraries in an environment +-without virtual memory management. This option implies @option{-fPIC}. ++@item -mno-pic ++@opindex mno-pic ++Do not emit position-independent code (will break dynamic linking.) + +-@item -mno-id-shared-library +-@opindex mno-id-shared-library +-Generate code that doesn't assume ID based shared libraries are being used. +-This is the default. ++@item -masm-addr-pseudos ++@opindex masm-addr-pseudos ++Use assembler pseudo-instructions lda.w and call for handling direct ++addresses. (Enabled by default) ++ ++@item -mno-asm-addr-pseudos ++@opindex mno-asm-addr-pseudos ++Do not use assembler pseudo-instructions lda.w and call for handling direct addresses. ++ ++@item -mpart=@var{part} ++@opindex mpart ++Generate code for the specified part. Permissible parts are: @samp{ap7000}, ++@samp{ap7010},@samp{ap7020},@samp{uc3a0256}, @samp{uc3a0512}, ++@samp{uc3a1128}, @samp{uc3a1256}, @samp{uc3a1512}. + +-@item -mshared-library-id=n +-@opindex mshared-library-id +-Specified the identification number of the ID based shared library being +-compiled. Specifying a value of 0 will generate more compact code, specifying +-other values will force the allocation of that number to the current +-library but is no more space or time efficient than omitting this option. ++@item -mcpu=@var{cpu-type} ++@opindex mcpu ++Same as -mpart. Obsolete. ++ ++@item -march=@var{arch} ++@opindex march ++Generate code for the specified architecture. Permissible architectures are: ++@samp{ap} and @samp{uc}. ++ ++@item -mfast-float ++@opindex mfast-float ++Enable fast floating-point library that does not conform to ieee but is still good enough ++for most applications. The fast floating-point library does not round to the nearest even ++but away from zero. Enabled by default if the -funsafe-math-optimizations switch is specified. + +-@item -mlong-calls +-@itemx -mno-long-calls +-@opindex mlong-calls +-@opindex mno-long-calls +-Tells the compiler to perform function calls by first loading the +-address of the function into a register and then performing a subroutine +-call on this register. This switch is needed if the target function +-will lie outside of the 24 bit addressing range of the offset based +-version of subroutine call instruction. + +-This feature is not enabled by default. Specifying +-@option{-mno-long-calls} will restore the default behavior. Note these +-switches have no effect on how the compiler generates code to handle +-function calls via function pointers. + @end table + + @node CRIS Options +@@ -11852,6 +11831,7 @@ conventions that adheres to the March 19 + Application Binary Interface, PowerPC processor supplement. This is the + default unless you configured GCC using @samp{powerpc-*-eabiaix}. + ++ + @item -mcall-sysv-eabi + @opindex mcall-sysv-eabi + Specify both @option{-mcall-sysv} and @option{-meabi} options. +@@ -14153,4 +14133,4 @@ You need to include the special files al + exist, because otherwise they won't get converted. + + @xref{Protoize Caveats}, for more information on how to use +-@code{protoize} successfully. ++@code{protoize} successfully. +\ No newline at end of file +diff -Nrup gcc-4.2.1/gcc/doc/md.texi gcc-4.2.1.atmel.1.3.2/gcc/doc/md.texi +--- gcc-4.2.1/gcc/doc/md.texi 2007-04-04 03:24:10.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/doc/md.texi 2007-09-28 10:32:30.000000000 +0200 +@@ -3,6 +3,7 @@ + @c This is part of the GCC manual. + @c For copying conditions, see the file gcc.texi. + ++ + @ifset INTERNALS + @node Machine Desc + @chapter Machine Descriptions +@@ -1681,6 +1682,58 @@ A memory reference suitable for iWMMXt l + A memory reference suitable for the ARMv4 ldrsb instruction. + @end table + ++@item AVR32 family---@file{avr32.h} ++@table @code ++@item f ++Floating-point registers (f0 to f15) ++ ++@item Ku@var{bits} ++Unsigned constant representable with @var{bits} number of bits (Must be ++two digits). I.e: An unsigned 8-bit constant is written as @samp{Ku08} ++ ++@item Ks@var{bits} ++Signed constant representable with @var{bits} number of bits (Must be ++two digits). I.e: A signed 12-bit constant is written as @samp{Ks12} ++ ++@item Is@var{bits} ++The negated range of a signed constant representable with @var{bits} ++number of bits. The same as @samp{Ks@var{bits}} with a negated range. ++This means that the constant must be in the range @math{-2^{bits-1}-1} to @math{2^{bits-1}} ++ ++@item G ++A single/double precision floating-point immediate or 64-bit integer ++immediate where the least and most significant words both can be ++loaded with a move instruction. That is the the integer form of the ++values in the least and most significant words both are in the range ++@math{-2^{20}} to @math{2^{20}-1}. ++ ++@item RKs@var{bits} ++A memory reference where the address consists of a base register ++plus a signed immediate displacement with range given by @samp{Ks@var{bits}} ++which has the same format as for the signed immediate integer constraint ++given above. ++ ++@item RKu@var{bits} ++A memory reference where the address consists of a base register ++plus an unsigned immediate displacement with range given by @samp{Ku@var{bits}} ++which has the same format as for the unsigned immediate integer constraint ++given above. ++ ++@item S ++A memory reference with an immediate or register offset ++ ++@item T ++A memory reference to a constant pool entry ++ ++@item W ++A valid operand for use in the @samp{lda.w} instruction macro when ++relaxing is enabled ++ ++@item Z ++A memory reference valid for coprocessor memory instructions ++ ++@end table ++ + @item AVR family---@file{config/avr/constraints.md} + @table @code + @item l +diff -Nrup gcc-4.2.1/gcc/expmed.c gcc-4.2.1.atmel.1.3.2/gcc/expmed.c +--- gcc-4.2.1/gcc/expmed.c 2007-04-25 07:43:32.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/expmed.c 2007-09-28 10:33:09.000000000 +0200 +@@ -37,6 +37,7 @@ Software Foundation, 51 Franklin Street, + #include "real.h" + #include "recog.h" + #include "langhooks.h" ++#include "target.h" + + static void store_fixed_bit_field (rtx, unsigned HOST_WIDE_INT, + unsigned HOST_WIDE_INT, +@@ -455,9 +456,18 @@ store_bit_field (rtx str_rtx, unsigned H + ? ((GET_MODE_SIZE (fieldmode) >= UNITS_PER_WORD + || GET_MODE_SIZE (GET_MODE (op0)) == GET_MODE_SIZE (fieldmode)) + && byte_offset % GET_MODE_SIZE (fieldmode) == 0) +- : (! SLOW_UNALIGNED_ACCESS (fieldmode, MEM_ALIGN (op0)) +- || (offset * BITS_PER_UNIT % bitsize == 0 +- && MEM_ALIGN (op0) % GET_MODE_BITSIZE (fieldmode) == 0)))) ++ : ( ++ ++ /* NB! Added for AVR32, and I think this should be true for ++ all targets not using narrow volatile bitfields. If the ++ bitfield is volatile then we need to perform an access ++ consistent with the container type. */ ++ !(MEM_VOLATILE_P (op0) ++ && GET_MODE_BITSIZE (GET_MODE (op0)) != bitsize ++ && !targetm.narrow_volatile_bitfield ()) ++ && (! SLOW_UNALIGNED_ACCESS (fieldmode, MEM_ALIGN (op0)) ++ || (offset * BITS_PER_UNIT % bitsize == 0 ++ && MEM_ALIGN (op0) % GET_MODE_BITSIZE (fieldmode) == 0))))) + { + if (MEM_P (op0)) + op0 = adjust_address (op0, fieldmode, offset); +@@ -1257,6 +1267,13 @@ extract_bit_field (rtx str_rtx, unsigned + && GET_MODE_SIZE (mode1) != 0 + && byte_offset % GET_MODE_SIZE (mode1) == 0) + || (MEM_P (op0) ++ /* NB! Added for AVR32, and I think this should be true for ++ all targets not using narrow volatile bitfields. If the ++ bitfield is volatile then we need to perform an access ++ consistent with the container type. */ ++ && !(MEM_VOLATILE_P (op0) ++ && GET_MODE_BITSIZE (GET_MODE (op0)) != bitsize ++ && !targetm.narrow_volatile_bitfield ()) + && (! SLOW_UNALIGNED_ACCESS (mode, MEM_ALIGN (op0)) + || (offset * BITS_PER_UNIT % bitsize == 0 + && MEM_ALIGN (op0) % bitsize == 0))))) +diff -Nrup gcc-4.2.1/gcc/expr.c gcc-4.2.1.atmel.1.3.2/gcc/expr.c +--- gcc-4.2.1/gcc/expr.c 2007-04-23 03:57:48.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/expr.c 2007-09-28 10:33:08.000000000 +0200 +@@ -3520,18 +3520,19 @@ emit_single_push_insn (enum machine_mode + } + else + { ++ emit_move_insn (stack_pointer_rtx, ++ expand_binop (Pmode, + #ifdef STACK_GROWS_DOWNWARD +- /* ??? This seems wrong if STACK_PUSH_CODE == POST_DEC. */ +- dest_addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, +- GEN_INT (-(HOST_WIDE_INT) rounded_size)); ++ sub_optab, + #else +- /* ??? This seems wrong if STACK_PUSH_CODE == POST_INC. */ +- dest_addr = gen_rtx_PLUS (Pmode, stack_pointer_rtx, +- GEN_INT (rounded_size)); ++ add_optab, + #endif +- dest_addr = gen_rtx_PRE_MODIFY (Pmode, stack_pointer_rtx, dest_addr); ++ stack_pointer_rtx, ++ GEN_INT (rounded_size), ++ NULL_RTX, 0, OPTAB_LIB_WIDEN)); ++ dest_addr = stack_pointer_rtx; + } +- ++ + dest = gen_rtx_MEM (mode, dest_addr); + + if (type != 0) +@@ -5510,7 +5511,16 @@ store_field (rtx target, HOST_WIDE_INT b + is a bit field, we cannot use addressing to access it. + Use bit-field techniques or SUBREG to store in it. */ + +- if (mode == VOIDmode ++ if ( ++ /* NB! Added for AVR32, and I think this should be true for ++ all targets not using narrow volatile bitfields. If the ++ bitfield is volatile then we need to perform an access ++ consistent with the container type. */ ++ (MEM_P (target) ++ && MEM_VOLATILE_P (target) ++ && GET_MODE_BITSIZE (GET_MODE (target)) > bitsize ++ && !targetm.narrow_volatile_bitfield ()) ++ || mode == VOIDmode + || (mode != BLKmode && ! direct_store[(int) mode] + && GET_MODE_CLASS (mode) != MODE_COMPLEX_INT + && GET_MODE_CLASS (mode) != MODE_COMPLEX_FLOAT) +@@ -7512,7 +7522,16 @@ expand_expr_real_1 (tree exp, rtx target + by doing the extract into an object as wide as the field + (which we know to be the width of a basic mode), then + storing into memory, and changing the mode to BLKmode. */ +- if (mode1 == VOIDmode ++ if ( ++ /* NB! Added for AVR32, and I think this should be true for ++ all targets not using narrow volatile bitfields. If the ++ bitfield is volatile then we need to perform an access ++ consistent with the container type. */ ++ (MEM_P (op0) ++ && MEM_VOLATILE_P (op0) ++ && GET_MODE_BITSIZE (GET_MODE (op0)) > bitsize ++ && !targetm.narrow_volatile_bitfield ()) ++ || mode1 == VOIDmode + || REG_P (op0) || GET_CODE (op0) == SUBREG + || (mode1 != BLKmode && ! direct_load[(int) mode1] + && GET_MODE_CLASS (mode) != MODE_COMPLEX_INT +diff -Nrup gcc-4.2.1/gcc/function.c gcc-4.2.1.atmel.1.3.2/gcc/function.c +--- gcc-4.2.1/gcc/function.c 2007-07-19 05:25:32.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/function.c 2007-09-28 10:33:09.000000000 +0200 +@@ -2677,8 +2677,12 @@ assign_parm_setup_reg (struct assign_par + SET_DECL_RTL (parm, parmreg); + + /* Copy the value into the register. */ +- if (data->nominal_mode != data->passed_mode +- || promoted_nominal_mode != data->promoted_mode) ++ if ( (data->nominal_mode != data->passed_mode ++ /* Added for AVR32: If passed_mode is equal ++ to promoted nominal mode why should be convert? ++ The conversion should make no difference. */ ++ && data->passed_mode != promoted_nominal_mode) ++ || promoted_nominal_mode != data->promoted_mode) + { + int save_tree_used; + +diff -Nrup gcc-4.2.1/gcc/genemit.c gcc-4.2.1.atmel.1.3.2/gcc/genemit.c +--- gcc-4.2.1/gcc/genemit.c 2006-03-29 23:07:12.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/genemit.c 2007-09-28 10:33:08.000000000 +0200 +@@ -122,6 +122,24 @@ max_operand_vec (rtx insn, int arg) + } + + static void ++gen_vararg_prologue(int operands) ++{ ++ int i; ++ ++ if (operands > 1) ++ { ++ for (i = 1; i < operands; i++) ++ printf(" rtx operand%d ATTRIBUTE_UNUSED;\n", i); ++ ++ printf(" va_list args;\n\n"); ++ printf(" va_start(args, operand0);\n"); ++ for (i = 1; i < operands; i++) ++ printf(" operand%d = va_arg(args, rtx);\n", i); ++ printf(" va_end(args);\n\n"); ++ } ++} ++ ++static void + print_code (RTX_CODE code) + { + const char *p1; +@@ -406,18 +424,16 @@ gen_insn (rtx insn, int lineno) + fatal ("match_dup operand number has no match_operand"); + + /* Output the function name and argument declarations. */ +- printf ("rtx\ngen_%s (", XSTR (insn, 0)); ++ printf ("rtx\ngen_%s ", XSTR (insn, 0)); ++ + if (operands) +- for (i = 0; i < operands; i++) +- if (i) +- printf (",\n\trtx operand%d ATTRIBUTE_UNUSED", i); +- else +- printf ("rtx operand%d ATTRIBUTE_UNUSED", i); ++ printf("(rtx operand0 ATTRIBUTE_UNUSED, ...)\n"); + else +- printf ("void"); +- printf (")\n"); ++ printf("(void)\n"); + printf ("{\n"); + ++ gen_vararg_prologue(operands); ++ + /* Output code to construct and return the rtl for the instruction body. */ + + if (XVECLEN (insn, 1) == 1) +@@ -457,16 +473,12 @@ gen_expand (rtx expand) + operands = max_operand_vec (expand, 1); + + /* Output the function name and argument declarations. */ +- printf ("rtx\ngen_%s (", XSTR (expand, 0)); ++ printf ("rtx\ngen_%s ", XSTR (expand, 0)); + if (operands) +- for (i = 0; i < operands; i++) +- if (i) +- printf (",\n\trtx operand%d", i); +- else +- printf ("rtx operand%d", i); ++ printf("(rtx operand0 ATTRIBUTE_UNUSED, ...)\n"); + else +- printf ("void"); +- printf (")\n"); ++ printf("(void)\n"); ++ + printf ("{\n"); + + /* If we don't have any C code to write, only one insn is being written, +@@ -476,6 +488,8 @@ gen_expand (rtx expand) + && operands > max_dup_opno + && XVECLEN (expand, 1) == 1) + { ++ gen_vararg_prologue(operands); ++ + printf (" return "); + gen_exp (XVECEXP (expand, 1, 0), DEFINE_EXPAND, NULL); + printf (";\n}\n\n"); +@@ -489,6 +503,7 @@ gen_expand (rtx expand) + for (; i <= max_scratch_opno; i++) + printf (" rtx operand%d ATTRIBUTE_UNUSED;\n", i); + printf (" rtx _val = 0;\n"); ++ gen_vararg_prologue(operands); + printf (" start_sequence ();\n"); + + /* The fourth operand of DEFINE_EXPAND is some code to be executed +diff -Nrup gcc-4.2.1/gcc/genflags.c gcc-4.2.1.atmel.1.3.2/gcc/genflags.c +--- gcc-4.2.1/gcc/genflags.c 2006-01-23 16:15:12.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/genflags.c 2007-09-28 10:33:09.000000000 +0200 +@@ -128,7 +128,6 @@ static void + gen_proto (rtx insn) + { + int num = num_operands (insn); +- int i; + const char *name = XSTR (insn, 0); + int truth = maybe_eval_c_test (XSTR (insn, 2)); + +@@ -159,12 +158,7 @@ gen_proto (rtx insn) + if (num == 0) + fputs ("void", stdout); + else +- { +- for (i = 1; i < num; i++) +- fputs ("rtx, ", stdout); +- +- fputs ("rtx", stdout); +- } ++ fputs("rtx, ...", stdout); + + puts (");"); + +@@ -174,12 +168,7 @@ gen_proto (rtx insn) + { + printf ("static inline rtx\ngen_%s", name); + if (num > 0) +- { +- putchar ('('); +- for (i = 0; i < num-1; i++) +- printf ("rtx ARG_UNUSED (%c), ", 'a' + i); +- printf ("rtx ARG_UNUSED (%c))\n", 'a' + i); +- } ++ puts("(rtx ARG_UNUSED(a), ...)"); + else + puts ("(void)"); + puts ("{\n return 0;\n}"); +diff -Nrup gcc-4.2.1/gcc/genoutput.c gcc-4.2.1.atmel.1.3.2/gcc/genoutput.c +--- gcc-4.2.1/gcc/genoutput.c 2006-03-29 23:07:12.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/genoutput.c 2007-09-28 10:33:08.000000000 +0200 +@@ -387,7 +387,7 @@ output_insn_data (void) + } + + if (d->name && d->name[0] != '*') +- printf (" (insn_gen_fn) gen_%s,\n", d->name); ++ printf (" gen_%s,\n", d->name); + else + printf (" 0,\n"); + +diff -Nrup gcc-4.2.1/gcc/ifcvt.c gcc-4.2.1.atmel.1.3.2/gcc/ifcvt.c +--- gcc-4.2.1/gcc/ifcvt.c 2006-11-15 09:37:38.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/ifcvt.c 2007-09-28 10:33:08.000000000 +0200 +@@ -1051,7 +1051,11 @@ noce_try_addcc (struct noce_if_info *if_ + != UNKNOWN)) + { + rtx cond = if_info->cond; +- enum rtx_code code = reversed_comparison_code (cond, if_info->jump); ++ /* This generates wrong code for AVR32. The cond code need not be reversed ++ since the addmodecc patterns add if the condition is NOT met. */ ++ /* enum rtx_code code = reversed_comparison_code (cond, if_info->jump);*/ ++ enum rtx_code code = GET_CODE(cond); ++ + + /* First try to use addcc pattern. */ + if (general_operand (XEXP (cond, 0), VOIDmode) +diff -Nrup gcc-4.2.1/gcc/longlong.h gcc-4.2.1.atmel.1.3.2/gcc/longlong.h +--- gcc-4.2.1/gcc/longlong.h 2006-06-13 19:44:56.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/longlong.h 2007-09-28 10:33:08.000000000 +0200 +@@ -227,6 +227,39 @@ UDItype __umulsidi3 (USItype, USItype); + #define UDIV_TIME 100 + #endif /* __arm__ */ + ++#if defined (__avr32__) && W_TYPE_SIZE == 32 ++#define add_ssaaaa(sh, sl, ah, al, bh, bl) \ ++ __asm__ ("add\t%1, %4, %5\n\tadc\t%0, %2, %3" \ ++ : "=r" ((USItype) (sh)), \ ++ "=&r" ((USItype) (sl)) \ ++ : "r" ((USItype) (ah)), \ ++ "r" ((USItype) (bh)), \ ++ "r" ((USItype) (al)), \ ++ "r" ((USItype) (bl)) __CLOBBER_CC) ++#define sub_ddmmss(sh, sl, ah, al, bh, bl) \ ++ __asm__ ("sub\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ ++ : "=r" ((USItype) (sh)), \ ++ "=&r" ((USItype) (sl)) \ ++ : "r" ((USItype) (ah)), \ ++ "r" ((USItype) (bh)), \ ++ "r" ((USItype) (al)), \ ++ "r" ((USItype) (bl)) __CLOBBER_CC) ++ ++#define __umulsidi3(a,b) ((UDItype)(a) * (UDItype)(b)) ++ ++#define umul_ppmm(w1, w0, u, v) \ ++{ \ ++ DWunion __w; \ ++ __w.ll = __umulsidi3 (u, v); \ ++ w1 = __w.s.high; \ ++ w0 = __w.s.low; \ ++} ++ ++#define count_leading_zeros(COUNT,X) ((COUNT) = __builtin_clz (X)) ++#define count_trailing_zeros(COUNT,X) ((COUNT) = __builtin_ctz (X)) ++#define COUNT_LEADING_ZEROS_0 32 ++#endif ++ + #if defined (__hppa) && W_TYPE_SIZE == 32 + #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ + __asm__ ("add %4,%5,%1\n\taddc %2,%3,%0" \ +diff -Nrup gcc-4.2.1/gcc/optabs.h gcc-4.2.1.atmel.1.3.2/gcc/optabs.h +--- gcc-4.2.1/gcc/optabs.h 2006-01-19 11:24:00.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/optabs.h 2007-09-28 10:33:07.000000000 +0200 +@@ -432,7 +432,7 @@ extern enum insn_code reload_out_optab[N + extern GTY(()) optab code_to_optab[NUM_RTX_CODE + 1]; + + +-typedef rtx (*rtxfun) (rtx); ++typedef rtx (*rtxfun) (rtx, ...); + + /* Indexed by the rtx-code for a conditional (e.g. EQ, LT,...) + gives the gen_function to make a branch to test that condition. */ +diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c +--- gcc-4.2.1/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c 2005-02-19 20:48:02.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/cpp/mac-eol-at-eof.c 2007-05-07 14:24:46.000000000 +0200 +@@ -1 +1,3 @@ +-/* Test no newline at eof warning when Mac line ending is used*/ +/* { dg-do compile } */ +int main() { return 0; } +\ No newline at end of file ++/* Test no newline at eof warning when Mac line ending is used*/ ++/* { dg-do compile } */ ++int main() { return 0; } +diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-3.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-3.c +--- gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-3.c 2005-07-20 08:39:38.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-3.c 2007-09-28 10:31:43.000000000 +0200 +@@ -5,7 +5,7 @@ + Copyright (C) 2002 Free Software Foundation Inc. + Contributed by Hans-Peter Nilsson */ + +-/* { dg-do run { xfail arc-*-* avr-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ ++/* { dg-do run { xfail arc-*-* avr-*-* avr32-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ + /* { dg-options "-O2 -foptimize-sibling-calls" } */ + + /* The option -foptimize-sibling-calls is the default, but serves as +diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-4.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-4.c +--- gcc-4.2.1/gcc/testsuite/gcc.dg/sibcall-4.c 2005-07-20 08:39:38.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/sibcall-4.c 2007-09-28 10:31:43.000000000 +0200 +@@ -5,7 +5,7 @@ + Copyright (C) 2002 Free Software Foundation Inc. + Contributed by Hans-Peter Nilsson */ + +-/* { dg-do run { xfail arc-*-* avr-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ ++/* { dg-do run { xfail arc-*-* avr-*-* avr32-*-* c4x-*-* cris-*-* h8300-*-* hppa*64*-*-* m32r-*-* m68hc1?-*-* m681?-*-* m680*-*-* m68k-*-* mcore-*-* mn10300-*-* xstormy16-*-* v850*-*-* vax-*-* xtensa-*-* } } */ + /* { dg-options "-O2 -foptimize-sibling-calls" } */ + + /* The option -foptimize-sibling-calls is the default, but serves as +diff -Nrup gcc-4.2.1/gcc/testsuite/gcc.dg/trampoline-1.c gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/trampoline-1.c +--- gcc-4.2.1/gcc/testsuite/gcc.dg/trampoline-1.c 2004-08-03 10:22:26.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gcc.dg/trampoline-1.c 2007-09-28 10:31:43.000000000 +0200 +@@ -46,6 +46,8 @@ void foo (void) + + int main (void) + { ++#ifndef NO_TRAMPOLINES + foo (); ++#endif + return 0; + } +diff -Nrup gcc-4.2.1/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 +--- gcc-4.2.1/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 2005-05-29 18:03:43.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/gfortran.dg/char_pointer_assign.f90 2007-05-07 16:38:14.000000000 +0200 +@@ -1,4 +1,4 @@ +-! { dg-do run } ++! { dg-do run } + program char_pointer_assign + ! Test character pointer assignments, required + ! to fix PR18890 and PR21297 +@@ -8,7 +8,7 @@ program char_pointer_assign + character*4, target :: t2(4) =(/"lmno","lmno","lmno","lmno"/) + character*4 :: const + character*4, pointer :: c1, c3 +- character*4, pointer :: c2(:), c4(:) ++ character*4, pointer :: c2(:), c4(:) + allocate (c3, c4(4)) + ! Scalars first. + c3 = "lmno" ! pointer = constant +@@ -24,13 +24,13 @@ program char_pointer_assign + + ! Now arrays. + c4 = "lmno" ! pointer = constant +- t2 = c4 ! target = pointer +- c2 => t2 ! pointer =>target +- const = c2(1) ++ t2 = c4 ! target = pointer ++ c2 => t2 ! pointer =>target ++ const = c2(1) + const(2:3) ="nm" ! c2(:)(2:3) = "nm" is still broken + c2 = const + c4 = c2 ! pointer = pointer +- const = c4(1) ++ const = c4(1) + const(1:1) ="o" ! c4(:)(1:1) = "o" is still broken + const(4:4) ="l" ! c4(:)(4:4) = "l" is still broken + c4 = const +diff -Nrup gcc-4.2.1/gcc/testsuite/g++.old-deja/g++.pt/static11.C gcc-4.2.1.atmel.1.3.2/gcc/testsuite/g++.old-deja/g++.pt/static11.C +--- gcc-4.2.1/gcc/testsuite/g++.old-deja/g++.pt/static11.C 2006-02-22 10:05:07.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/gcc/testsuite/g++.old-deja/g++.pt/static11.C 2007-09-28 10:31:46.000000000 +0200 +@@ -2,7 +2,7 @@ + // in their dejagnu baseboard description) require that the status is + // final when exit is entered (or main returns), and not "overruled" by a + // destructor calling _exit. It's not really worth it to handle that. +-// { dg-do run { xfail mmix-knuth-mmixware xtensa-*-elf* arm*-*-elf arm*-*-eabi m68k-*-elf } } ++// { dg-do run { xfail mmix-knuth-mmixware xtensa-*-elf* avr32-*-elf arm*-*-elf arm*-*-eabi m68k-*-elf } } + + // Bug: g++ was failing to destroy C::a because it was using two + // different sentry variables for construction and destruction. + +diff -Nrup gcc-4.2.1/libstdc++-v3/acinclude.m4 gcc-4.2.1.atmel.1.3.2/libstdc++-v3/acinclude.m4 +--- gcc-4.2.1/libstdc++-v3/acinclude.m4 2007-06-29 01:02:05.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/acinclude.m4 2007-09-28 10:33:34.000000000 +0200 +@@ -125,15 +125,6 @@ AC_DEFUN([GLIBCXX_CONFIGURE], [ + ## other macros from doing the same. This should be automated.) -pme + need_libmath=no + +- # Check for uClibc since Linux platforms use different configuration +- # directories depending on the C library in use. +- AC_EGREP_CPP([_using_uclibc], [ +- #include +- #if __UCLIBC__ +- _using_uclibc +- #endif +- ], uclibc=yes, uclibc=no) +- + # Find platform-specific directories containing configuration info. + # Also possibly modify flags used elsewhere, as needed by the platform. + GLIBCXX_CHECK_HOST +@@ -1389,8 +1380,8 @@ AC_DEFUN([GLIBCXX_ENABLE_CLOCALE], [ + #endif + int main() + { +- const char __one[] = "�uglein Augmen"; +- const char __two[] = "�uglein"; ++ const char __one[] = "Äuglein Augmen"; ++ const char __two[] = "Äuglein"; + int i; + int j; + __locale_t loc; +diff -Nrup gcc-4.2.1/libstdc++-v3/config/os/gnu-linux/ctype_base.h gcc-4.2.1.atmel.1.3.2/libstdc++-v3/config/os/gnu-linux/ctype_base.h +--- gcc-4.2.1/libstdc++-v3/config/os/gnu-linux/ctype_base.h 2006-12-01 13:56:23.000000000 +0100 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/config/os/gnu-linux/ctype_base.h 2007-09-28 10:33:32.000000000 +0200 +@@ -31,6 +31,8 @@ + // + // ISO C++ 14882: 22.1 Locales + // ++#include ++#include + + /** @file ctype_base.h + * This is an internal header file, included by other library headers. +@@ -45,8 +47,12 @@ _GLIBCXX_BEGIN_NAMESPACE(std) + struct ctype_base + { + // Non-standard typedefs. +- typedef const int* __to_type; +- ++#ifdef __UCLIBC__ ++ typedef const __ctype_touplow_t* __to_type; ++#else ++ typedef const int* __to_type; ++#endif ++ + // NB: Offsets into ctype::_M_table force a particular size + // on the mask type. Because of this, we don't use an enum. + typedef unsigned short mask; +diff -Nrup gcc-4.2.1/libstdc++-v3/include/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/include/Makefile.in +--- gcc-4.2.1/libstdc++-v3/include/Makefile.in 2007-07-05 13:46:00.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/include/Makefile.in 2007-09-28 10:33:21.000000000 +0200 +@@ -36,6 +36,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ + $(top_srcdir)/fragment.am + subdir = include +diff -Nrup gcc-4.2.1/libstdc++-v3/libmath/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libmath/Makefile.in +--- gcc-4.2.1/libstdc++-v3/libmath/Makefile.in 2006-10-16 21:08:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libmath/Makefile.in 2007-09-28 10:33:16.000000000 +0200 +@@ -37,6 +37,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + subdir = libmath + DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in + ACLOCAL_M4 = $(top_srcdir)/aclocal.m4 +diff -Nrup gcc-4.2.1/libstdc++-v3/libsupc++/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libsupc++/Makefile.in +--- gcc-4.2.1/libstdc++-v3/libsupc++/Makefile.in 2006-10-16 21:08:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/libsupc++/Makefile.in 2007-09-28 10:33:21.000000000 +0200 +@@ -38,6 +38,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + DIST_COMMON = $(glibcxxinstall_HEADERS) $(srcdir)/Makefile.am \ + $(srcdir)/Makefile.in $(top_srcdir)/fragment.am + subdir = libsupc++ +diff -Nrup gcc-4.2.1/libstdc++-v3/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/Makefile.in +--- gcc-4.2.1/libstdc++-v3/Makefile.in 2006-10-16 21:08:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/Makefile.in 2007-09-28 10:33:34.000000000 +0200 +@@ -36,6 +36,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + DIST_COMMON = README $(am__configure_deps) $(srcdir)/../config.guess \ + $(srcdir)/../config.sub $(srcdir)/../install-sh \ + $(srcdir)/../ltmain.sh $(srcdir)/../missing \ +diff -Nrup gcc-4.2.1/libstdc++-v3/po/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/po/Makefile.in +--- gcc-4.2.1/libstdc++-v3/po/Makefile.in 2006-10-16 21:08:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/po/Makefile.in 2007-09-28 10:33:34.000000000 +0200 +@@ -36,6 +36,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ + $(top_srcdir)/fragment.am + subdir = po +diff -Nrup gcc-4.2.1/libstdc++-v3/src/Makefile.in gcc-4.2.1.atmel.1.3.2/libstdc++-v3/src/Makefile.in +--- gcc-4.2.1/libstdc++-v3/src/Makefile.in 2006-10-16 21:08:22.000000000 +0200 ++++ gcc-4.2.1.atmel.1.3.2/libstdc++-v3/src/Makefile.in 2007-09-28 10:33:21.000000000 +0200 +@@ -36,6 +36,7 @@ POST_UNINSTALL = : + build_triplet = @build@ + host_triplet = @host@ + target_triplet = @target@ ++LIBOBJDIR = + DIST_COMMON = $(srcdir)/Makefile.am $(srcdir)/Makefile.in \ + $(top_srcdir)/fragment.am + subdir = src diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/901-avr32-add-mno-pic.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/901-avr32-add-mno-pic.patch new file mode 100644 index 0000000000..c21450a9ad --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/901-avr32-add-mno-pic.patch @@ -0,0 +1,25 @@ +Index: gcc/gcc/config/avr32/avr32.opt +=================================================================== +--- gcc/gcc/config/avr32/avr32.opt (revision 31583) ++++ gcc/gcc/config/avr32/avr32.opt (working copy) +@@ -71,3 +71,6 @@ + Target Report Var(avr32_imm_in_const_pool) Init(-1) + Put large immediates in constant pool. This is enabled by default for archs with insn-cache. + ++mno-pic ++Target Report RejectNegative Mask(NO_PIC) ++Do not generate position-independent code. +Index: gcc/gcc/config/avr32/avr32.c +=================================================================== +--- gcc/gcc/config/avr32/avr32.c (revision 31583) ++++ gcc/gcc/config/avr32/avr32.c (working copy) +@@ -230,6 +230,9 @@ + avr32_imm_in_const_pool = 0; + } + ++ if (TARGET_NO_PIC) ++ flag_pic = 0; ++ + avr32_add_gc_roots (); + } + diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/902-avr32-fix-pic-redefinition-warning.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/902-avr32-fix-pic-redefinition-warning.patch new file mode 100644 index 0000000000..ae9fa5d188 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/902-avr32-fix-pic-redefinition-warning.patch @@ -0,0 +1,16 @@ +Index: gcc/gcc/config/avr32/linux-elf.h +=================================================================== +--- gcc/gcc/config/avr32/linux-elf.h (revision 31583) ++++ gcc/gcc/config/avr32/linux-elf.h (working copy) +@@ -122,11 +122,6 @@ + builtin_define ("__AVR32_HAS_BRANCH_PRED__"); \ + if (TARGET_FAST_FLOAT) \ + builtin_define ("__AVR32_FAST_FLOAT__"); \ +- if (flag_pic) \ +- { \ +- builtin_define ("__PIC__"); \ +- builtin_define ("__pic__"); \ +- } \ + } \ + while (0) + diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/903-avr32-fix-expanding-builtin-xchg.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/903-avr32-fix-expanding-builtin-xchg.patch new file mode 100644 index 0000000000..4a4ac60932 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/903-avr32-fix-expanding-builtin-xchg.patch @@ -0,0 +1,12 @@ +--- gcc-4.2.1.orig/gcc/config/avr32/avr32.c 2007-10-05 07:52:51.000000000 +0200 ++++ gcc-4.2.1/gcc/config/avr32/avr32.c 2007-10-05 07:54:49.000000000 +0200 +@@ -1388,7 +1388,8 @@ avr32_expand_builtin (tree exp, + op1 = copy_to_mode_reg (mode1, op1); + } + +- op0 = gen_rtx_MEM (SImode, op0); ++ op0 = force_reg (GET_MODE (op0), op0); ++ op0 = gen_rtx_MEM (GET_MODE (op0), op0); + if (!(*insn_data[icode].operand[1].predicate) (op0, mode0)) + { + error diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/905-avr32-fix-folding-machine-reorg-optimizations.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/905-avr32-fix-folding-machine-reorg-optimizations.patch new file mode 100644 index 0000000000..7a160d6e95 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/905-avr32-fix-folding-machine-reorg-optimizations.patch @@ -0,0 +1,48 @@ +Index: a/gcc/config/avr32/avr32.c +=================================================================== +--- a/gcc/config/avr32/avr32.c (revision 31997) ++++ b/gcc/config/avr32/avr32.c (working copy) +@@ -6323,25 +6323,26 @@ + /* If used any other place than as a pointer or as the + destination register we failed */ + if (!(single_set (scan) +- && GET_CODE (PATTERN (scan)) == SET +- && ((MEM_P (SET_DEST (PATTERN (scan))) +- && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) +- && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == +- REGNO (reg)) || (MEM_P (SET_SRC (PATTERN (scan))) +- && +- REG_P (XEXP +- (SET_SRC (PATTERN (scan)), +- 0)) +- && +- REGNO (XEXP +- (SET_SRC (PATTERN (scan)), +- 0)) == REGNO (reg)))) +- && !(GET_CODE (PATTERN (scan)) == SET +- && REG_P (SET_DEST (PATTERN (scan))) +- && !regno_use_in (REGNO (reg), +- SET_SRC (PATTERN (scan))))) ++ && GET_CODE (PATTERN (scan)) == SET ++ && ((MEM_P (SET_DEST (PATTERN (scan))) ++ && REG_P (XEXP (SET_DEST (PATTERN (scan)), 0)) ++ && REGNO (XEXP (SET_DEST (PATTERN (scan)), 0)) == REGNO (reg)) ++ || (MEM_P (SET_SRC (PATTERN (scan))) ++ && REG_P (XEXP (SET_SRC (PATTERN (scan)), 0)) ++ && REGNO (XEXP ++ (SET_SRC (PATTERN (scan)), 0)) == REGNO (reg)))) ++ && !(GET_CODE (PATTERN (scan)) == SET ++ && REG_P (SET_DEST (PATTERN (scan))) ++ && !regno_use_in (REGNO (reg), ++ SET_SRC (PATTERN (scan))))) + break; + ++ /* We cannot replace the pointer in TImode insns ++ as these has a differene addressing mode than the other ++ memory insns. */ ++ if ( GET_MODE (SET_DEST (PATTERN (scan))) == TImode ) ++ break; ++ + /* Check if register is dead or set in this insn */ + if (dead_or_set_p (scan, reg)) + { diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/906-avr32-use-rjmp-instead-of-got-when-jumping.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/906-avr32-use-rjmp-instead-of-got-when-jumping.patch new file mode 100644 index 0000000000..5d3c8d18a3 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/906-avr32-use-rjmp-instead-of-got-when-jumping.patch @@ -0,0 +1,85 @@ +Index: a/gcc/config/avr32/avr32.c +=================================================================== +--- a/gcc/config/avr32/avr32.c (revision 32101) ++++ b/gcc/config/avr32/avr32.c (working copy) +@@ -695,8 +695,7 @@ + + + if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") +- || vcall_offset +- || flag_pic) ++ || vcall_offset) + { + fputs ("\tpushm\tlr\n", file); + } +@@ -728,47 +727,23 @@ + } + + +- if ( (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") +- || vcall_offset) +- && !flag_pic ) ++ if (!avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21") ++ || vcall_offset) + { + fputs ("\tpopm\tlr\n", file); + } +- +- if (flag_pic) +- { +- /* Load the got into lr and then load the pointer +- to the function from the got and put it on the stack. +- We can then call the function and restore lr by issuing +- a doubleword load from the stack. We do not use a popm/ldm +- since it will be treated as a return and might need a flushing +- of the return-stack if available. */ +- rtx label = gen_label_rtx (); +- /* Load the got. */ +- fputs ("\tlddpc\tlr, 0f\n", file); +- (*targetm.asm_out.internal_label) (file, "L", +- CODE_LABEL_NUMBER (label)); +- fputs ("\trsub\tlr, pc\n", file); +- /* Load the function pointer. */ +- fputs ("\tld.w\tlr, lr[", file); +- assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); +- fputs ("@got]\n", file); +- /* Push the function pointer on the stack.*/ +- fputs ("\tpushm\tlr\n", file); +- /* Restore the old lr value and load the function pointer into +- pc. */ +- fputs ("\tld.d\tlr,sp++\n", file); +- fprintf (file, "\t.align 2\n"); +- fprintf (file, "0:\t.long\t.L%d - _GLOBAL_OFFSET_TABLE_\n", CODE_LABEL_NUMBER (label)); +- } +- else +- { +- fprintf (file, "\tlddpc\tpc, 0f\n"); +- fprintf (file, "\t.align 2\n"); +- fputs ("0:\t.long\t", file); +- assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); +- fputc ('\n', file); +- } ++ ++ /* Jump to the function. We assume that we can use an rjmp since the ++ function to jump to is local and probably not too far away from ++ the thunk. If this assumption proves to be wrong we could implement ++ this jump by calculating the offset between the jump source and destination ++ and put this in the constant pool and then perform an add to pc. ++ This would also be legitimate PIC code. But for now we hope that an rjmp ++ will be sufficient... ++ */ ++ fputs ("\trjmp\t", file); ++ assemble_name (file, XSTR (XEXP (DECL_RTL (function), 0), 0)); ++ fputc ('\n', file); + } + + /* Implements target hook vector_mode_supported. */ +@@ -1742,7 +1717,7 @@ + if (TREE_CODE (*node) != FUNCTION_DECL) + { + warning ("`%s' attribute only applies to functions", +- IDENTIFIER_POINTER (name)); ++ IDENTIFIER_POINTER (name)); + *no_add_attrs = true; + } + /* FIXME: the argument if any is checked for type attributes; should it diff --git a/toolchain/gcc/4.2.1/avr32_patches_break_x86/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch b/toolchain/gcc/4.2.1/avr32_patches_break_x86/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch new file mode 100644 index 0000000000..63ead2be79 --- /dev/null +++ b/toolchain/gcc/4.2.1/avr32_patches_break_x86/908-avr32-fix-invalid-assembler-code-on-64bit-hosts.patch @@ -0,0 +1,14 @@ +diff -Nrup gcc-4.2.1/gcc/config/avr32/avr32.c gcc-4.2.1-new/gcc/config/avr32/avr32.c +--- gcc-4.2.1/gcc/config/avr32/avr32.c 2007-11-15 18:20:25.000000000 +0100 ++++ gcc-4.2.1-new/gcc/config/avr32/avr32.c 2007-11-16 13:34:39.000000000 +0100 +@@ -705,8 +705,8 @@ avr32_output_mi_thunk (FILE * file, + { + if (avr32_const_ok_for_constraint_p (mi_delta, 'I', "Is21")) + { +- fprintf (file, "\tsub\t%s, -0x%x\n", reg_names[this_regno], +- mi_delta); ++ fprintf (file, "\tsub\t%s, %d\n", reg_names[this_regno], ++ -mi_delta); + } + else + {