diff options
Diffstat (limited to 'gcc-4.8.1/gcc/config/aarch64/aarch64.c')
-rw-r--r-- | gcc-4.8.1/gcc/config/aarch64/aarch64.c | 8007 |
1 files changed, 0 insertions, 8007 deletions
diff --git a/gcc-4.8.1/gcc/config/aarch64/aarch64.c b/gcc-4.8.1/gcc/config/aarch64/aarch64.c deleted file mode 100644 index 4166782b8..000000000 --- a/gcc-4.8.1/gcc/config/aarch64/aarch64.c +++ /dev/null @@ -1,8007 +0,0 @@ -/* Machine description for AArch64 architecture. - Copyright (C) 2009-2013 Free Software Foundation, Inc. - Contributed by ARM Ltd. - - This file is part of GCC. - - GCC is free software; you can redistribute it and/or modify it - under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 3, or (at your option) - any later version. - - GCC is distributed in the hope that it will be useful, but - WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - General Public License for more details. - - You should have received a copy of the GNU General Public License - along with GCC; see the file COPYING3. If not see - <http://www.gnu.org/licenses/>. */ - -#include "config.h" -#include "system.h" -#include "coretypes.h" -#include "tm.h" -#include "insn-codes.h" -#include "rtl.h" -#include "insn-attr.h" -#include "tree.h" -#include "regs.h" -#include "df.h" -#include "hard-reg-set.h" -#include "output.h" -#include "expr.h" -#include "reload.h" -#include "toplev.h" -#include "target.h" -#include "target-def.h" -#include "targhooks.h" -#include "ggc.h" -#include "function.h" -#include "tm_p.h" -#include "recog.h" -#include "langhooks.h" -#include "diagnostic-core.h" -#include "gimple.h" -#include "optabs.h" -#include "dwarf2.h" - -/* Classifies an address. - - ADDRESS_REG_IMM - A simple base register plus immediate offset. - - ADDRESS_REG_WB - A base register indexed by immediate offset with writeback. - - ADDRESS_REG_REG - A base register indexed by (optionally scaled) register. - - ADDRESS_REG_UXTW - A base register indexed by (optionally scaled) zero-extended register. - - ADDRESS_REG_SXTW - A base register indexed by (optionally scaled) sign-extended register. - - ADDRESS_LO_SUM - A LO_SUM rtx with a base register and "LO12" symbol relocation. - - ADDRESS_SYMBOLIC: - A constant symbolic address, in pc-relative literal pool. */ - -enum aarch64_address_type { - ADDRESS_REG_IMM, - ADDRESS_REG_WB, - ADDRESS_REG_REG, - ADDRESS_REG_UXTW, - ADDRESS_REG_SXTW, - ADDRESS_LO_SUM, - ADDRESS_SYMBOLIC -}; - -struct aarch64_address_info { - enum aarch64_address_type type; - rtx base; - rtx offset; - int shift; - enum aarch64_symbol_type symbol_type; -}; - -/* The current code model. */ -enum aarch64_code_model aarch64_cmodel; - -#ifdef HAVE_AS_TLS -#undef TARGET_HAVE_TLS -#define TARGET_HAVE_TLS 1 -#endif - -static bool aarch64_composite_type_p (const_tree, enum machine_mode); -static bool aarch64_vfp_is_call_or_return_candidate (enum machine_mode, - const_tree, - enum machine_mode *, int *, - bool *); -static void aarch64_elf_asm_constructor (rtx, int) ATTRIBUTE_UNUSED; -static void aarch64_elf_asm_destructor (rtx, int) ATTRIBUTE_UNUSED; -static void aarch64_override_options_after_change (void); -static int aarch64_simd_valid_immediate (rtx, enum machine_mode, int, rtx *, - int *, unsigned char *, int *, int *); -static bool aarch64_vector_mode_supported_p (enum machine_mode); -static unsigned bit_count (unsigned HOST_WIDE_INT); -static bool aarch64_const_vec_all_same_int_p (rtx, - HOST_WIDE_INT, HOST_WIDE_INT); - -static bool aarch64_vectorize_vec_perm_const_ok (enum machine_mode vmode, - const unsigned char *sel); - -/* The processor for which instructions should be scheduled. */ -enum aarch64_processor aarch64_tune = generic; - -/* The current tuning set. */ -const struct tune_params *aarch64_tune_params; - -/* Mask to specify which instructions we are allowed to generate. */ -unsigned long aarch64_isa_flags = 0; - -/* Mask to specify which instruction scheduling options should be used. */ -unsigned long aarch64_tune_flags = 0; - -/* Tuning parameters. */ - -#if HAVE_DESIGNATED_INITIALIZERS -#define NAMED_PARAM(NAME, VAL) .NAME = (VAL) -#else -#define NAMED_PARAM(NAME, VAL) (VAL) -#endif - -#if HAVE_DESIGNATED_INITIALIZERS && GCC_VERSION >= 2007 -__extension__ -#endif -static const struct cpu_rtx_cost_table generic_rtx_cost_table = -{ - NAMED_PARAM (memory_load, COSTS_N_INSNS (1)), - NAMED_PARAM (memory_store, COSTS_N_INSNS (0)), - NAMED_PARAM (register_shift, COSTS_N_INSNS (1)), - NAMED_PARAM (int_divide, COSTS_N_INSNS (6)), - NAMED_PARAM (float_divide, COSTS_N_INSNS (2)), - NAMED_PARAM (double_divide, COSTS_N_INSNS (6)), - NAMED_PARAM (int_multiply, COSTS_N_INSNS (1)), - NAMED_PARAM (int_multiply_extend, COSTS_N_INSNS (1)), - NAMED_PARAM (int_multiply_add, COSTS_N_INSNS (1)), - NAMED_PARAM (int_multiply_extend_add, COSTS_N_INSNS (1)), - NAMED_PARAM (float_multiply, COSTS_N_INSNS (0)), - NAMED_PARAM (double_multiply, COSTS_N_INSNS (1)) -}; - -#if HAVE_DESIGNATED_INITIALIZERS && GCC_VERSION >= 2007 -__extension__ -#endif -static const struct cpu_addrcost_table generic_addrcost_table = -{ - NAMED_PARAM (pre_modify, 0), - NAMED_PARAM (post_modify, 0), - NAMED_PARAM (register_offset, 0), - NAMED_PARAM (register_extend, 0), - NAMED_PARAM (imm_offset, 0) -}; - -#if HAVE_DESIGNATED_INITIALIZERS && GCC_VERSION >= 2007 -__extension__ -#endif -static const struct cpu_regmove_cost generic_regmove_cost = -{ - NAMED_PARAM (GP2GP, 1), - NAMED_PARAM (GP2FP, 2), - NAMED_PARAM (FP2GP, 2), - /* We currently do not provide direct support for TFmode Q->Q move. - Therefore we need to raise the cost above 2 in order to have - reload handle the situation. */ - NAMED_PARAM (FP2FP, 4) -}; - -#if HAVE_DESIGNATED_INITIALIZERS && GCC_VERSION >= 2007 -__extension__ -#endif -static const struct tune_params generic_tunings = -{ - &generic_rtx_cost_table, - &generic_addrcost_table, - &generic_regmove_cost, - NAMED_PARAM (memmov_cost, 4) -}; - -/* A processor implementing AArch64. */ -struct processor -{ - const char *const name; - enum aarch64_processor core; - const char *arch; - const unsigned long flags; - const struct tune_params *const tune; -}; - -/* Processor cores implementing AArch64. */ -static const struct processor all_cores[] = -{ -#define AARCH64_CORE(NAME, IDENT, ARCH, FLAGS, COSTS) \ - {NAME, IDENT, #ARCH, FLAGS | AARCH64_FL_FOR_ARCH##ARCH, &COSTS##_tunings}, -#include "aarch64-cores.def" -#undef AARCH64_CORE - {"generic", generic, "8", AARCH64_FL_FPSIMD | AARCH64_FL_FOR_ARCH8, &generic_tunings}, - {NULL, aarch64_none, NULL, 0, NULL} -}; - -/* Architectures implementing AArch64. */ -static const struct processor all_architectures[] = -{ -#define AARCH64_ARCH(NAME, CORE, ARCH, FLAGS) \ - {NAME, CORE, #ARCH, FLAGS, NULL}, -#include "aarch64-arches.def" -#undef AARCH64_ARCH - {"generic", generic, "8", AARCH64_FL_FOR_ARCH8, NULL}, - {NULL, aarch64_none, NULL, 0, NULL} -}; - -/* Target specification. These are populated as commandline arguments - are processed, or NULL if not specified. */ -static const struct processor *selected_arch; -static const struct processor *selected_cpu; -static const struct processor *selected_tune; - -#define AARCH64_CPU_DEFAULT_FLAGS ((selected_cpu) ? selected_cpu->flags : 0) - -/* An ISA extension in the co-processor and main instruction set space. */ -struct aarch64_option_extension -{ - const char *const name; - const unsigned long flags_on; - const unsigned long flags_off; -}; - -/* ISA extensions in AArch64. */ -static const struct aarch64_option_extension all_extensions[] = -{ -#define AARCH64_OPT_EXTENSION(NAME, FLAGS_ON, FLAGS_OFF) \ - {NAME, FLAGS_ON, FLAGS_OFF}, -#include "aarch64-option-extensions.def" -#undef AARCH64_OPT_EXTENSION - {NULL, 0, 0} -}; - -/* Used to track the size of an address when generating a pre/post - increment address. */ -static enum machine_mode aarch64_memory_reference_mode; - -/* Used to force GTY into this file. */ -static GTY(()) int gty_dummy; - -/* A table of valid AArch64 "bitmask immediate" values for - logical instructions. */ - -#define AARCH64_NUM_BITMASKS 5334 -static unsigned HOST_WIDE_INT aarch64_bitmasks[AARCH64_NUM_BITMASKS]; - -/* Did we set flag_omit_frame_pointer just so - aarch64_frame_pointer_required would be called? */ -static bool faked_omit_frame_pointer; - -typedef enum aarch64_cond_code -{ - AARCH64_EQ = 0, AARCH64_NE, AARCH64_CS, AARCH64_CC, AARCH64_MI, AARCH64_PL, - AARCH64_VS, AARCH64_VC, AARCH64_HI, AARCH64_LS, AARCH64_GE, AARCH64_LT, - AARCH64_GT, AARCH64_LE, AARCH64_AL, AARCH64_NV -} -aarch64_cc; - -#define AARCH64_INVERSE_CONDITION_CODE(X) ((aarch64_cc) (((int) X) ^ 1)) - -/* The condition codes of the processor, and the inverse function. */ -static const char * const aarch64_condition_codes[] = -{ - "eq", "ne", "cs", "cc", "mi", "pl", "vs", "vc", - "hi", "ls", "ge", "lt", "gt", "le", "al", "nv" -}; - -/* Provide a mapping from gcc register numbers to dwarf register numbers. */ -unsigned -aarch64_dbx_register_number (unsigned regno) -{ - if (GP_REGNUM_P (regno)) - return AARCH64_DWARF_R0 + regno - R0_REGNUM; - else if (regno == SP_REGNUM) - return AARCH64_DWARF_SP; - else if (FP_REGNUM_P (regno)) - return AARCH64_DWARF_V0 + regno - V0_REGNUM; - - /* Return values >= DWARF_FRAME_REGISTERS indicate that there is no - equivalent DWARF register. */ - return DWARF_FRAME_REGISTERS; -} - -/* Return TRUE if MODE is any of the large INT modes. */ -static bool -aarch64_vect_struct_mode_p (enum machine_mode mode) -{ - return mode == OImode || mode == CImode || mode == XImode; -} - -/* Return TRUE if MODE is any of the vector modes. */ -static bool -aarch64_vector_mode_p (enum machine_mode mode) -{ - return aarch64_vector_mode_supported_p (mode) - || aarch64_vect_struct_mode_p (mode); -} - -/* Implement target hook TARGET_ARRAY_MODE_SUPPORTED_P. */ -static bool -aarch64_array_mode_supported_p (enum machine_mode mode, - unsigned HOST_WIDE_INT nelems) -{ - if (TARGET_SIMD - && AARCH64_VALID_SIMD_QREG_MODE (mode) - && (nelems >= 2 && nelems <= 4)) - return true; - - return false; -} - -/* Implement HARD_REGNO_NREGS. */ - -int -aarch64_hard_regno_nregs (unsigned regno, enum machine_mode mode) -{ - switch (aarch64_regno_regclass (regno)) - { - case FP_REGS: - case FP_LO_REGS: - return (GET_MODE_SIZE (mode) + UNITS_PER_VREG - 1) / UNITS_PER_VREG; - default: - return (GET_MODE_SIZE (mode) + UNITS_PER_WORD - 1) / UNITS_PER_WORD; - } - gcc_unreachable (); -} - -/* Implement HARD_REGNO_MODE_OK. */ - -int -aarch64_hard_regno_mode_ok (unsigned regno, enum machine_mode mode) -{ - if (GET_MODE_CLASS (mode) == MODE_CC) - return regno == CC_REGNUM; - - if (regno == SP_REGNUM || regno == FRAME_POINTER_REGNUM - || regno == ARG_POINTER_REGNUM) - return mode == Pmode; - - if (GP_REGNUM_P (regno) && ! aarch64_vect_struct_mode_p (mode)) - return 1; - - if (FP_REGNUM_P (regno)) - { - if (aarch64_vect_struct_mode_p (mode)) - return - (regno + aarch64_hard_regno_nregs (regno, mode) - 1) <= V31_REGNUM; - else - return 1; - } - - return 0; -} - -/* Return true if calls to DECL should be treated as - long-calls (ie called via a register). */ -static bool -aarch64_decl_is_long_call_p (const_tree decl ATTRIBUTE_UNUSED) -{ - return false; -} - -/* Return true if calls to symbol-ref SYM should be treated as - long-calls (ie called via a register). */ -bool -aarch64_is_long_call_p (rtx sym) -{ - return aarch64_decl_is_long_call_p (SYMBOL_REF_DECL (sym)); -} - -/* Return true if the offsets to a zero/sign-extract operation - represent an expression that matches an extend operation. The - operands represent the paramters from - - (extract (mult (reg) (mult_imm)) (extract_imm) (const_int 0)). */ -bool -aarch64_is_extend_from_extract (enum machine_mode mode, rtx mult_imm, - rtx extract_imm) -{ - HOST_WIDE_INT mult_val, extract_val; - - if (! CONST_INT_P (mult_imm) || ! CONST_INT_P (extract_imm)) - return false; - - mult_val = INTVAL (mult_imm); - extract_val = INTVAL (extract_imm); - - if (extract_val > 8 - && extract_val < GET_MODE_BITSIZE (mode) - && exact_log2 (extract_val & ~7) > 0 - && (extract_val & 7) <= 4 - && mult_val == (1 << (extract_val & 7))) - return true; - - return false; -} - -/* Emit an insn that's a simple single-set. Both the operands must be - known to be valid. */ -inline static rtx -emit_set_insn (rtx x, rtx y) -{ - return emit_insn (gen_rtx_SET (VOIDmode, x, y)); -} - -/* X and Y are two things to compare using CODE. Emit the compare insn and - return the rtx for register 0 in the proper mode. */ -rtx -aarch64_gen_compare_reg (RTX_CODE code, rtx x, rtx y) -{ - enum machine_mode mode = SELECT_CC_MODE (code, x, y); - rtx cc_reg = gen_rtx_REG (mode, CC_REGNUM); - - emit_set_insn (cc_reg, gen_rtx_COMPARE (mode, x, y)); - return cc_reg; -} - -/* Build the SYMBOL_REF for __tls_get_addr. */ - -static GTY(()) rtx tls_get_addr_libfunc; - -rtx -aarch64_tls_get_addr (void) -{ - if (!tls_get_addr_libfunc) - tls_get_addr_libfunc = init_one_libfunc ("__tls_get_addr"); - return tls_get_addr_libfunc; -} - -/* Return the TLS model to use for ADDR. */ - -static enum tls_model -tls_symbolic_operand_type (rtx addr) -{ - enum tls_model tls_kind = TLS_MODEL_NONE; - rtx sym, addend; - - if (GET_CODE (addr) == CONST) - { - split_const (addr, &sym, &addend); - if (GET_CODE (sym) == SYMBOL_REF) - tls_kind = SYMBOL_REF_TLS_MODEL (sym); - } - else if (GET_CODE (addr) == SYMBOL_REF) - tls_kind = SYMBOL_REF_TLS_MODEL (addr); - - return tls_kind; -} - -/* We'll allow lo_sum's in addresses in our legitimate addresses - so that combine would take care of combining addresses where - necessary, but for generation purposes, we'll generate the address - as : - RTL Absolute - tmp = hi (symbol_ref); adrp x1, foo - dest = lo_sum (tmp, symbol_ref); add dest, x1, :lo_12:foo - nop - - PIC TLS - adrp x1, :got:foo adrp tmp, :tlsgd:foo - ldr x1, [:got_lo12:foo] add dest, tmp, :tlsgd_lo12:foo - bl __tls_get_addr - nop - - Load TLS symbol, depending on TLS mechanism and TLS access model. - - Global Dynamic - Traditional TLS: - adrp tmp, :tlsgd:imm - add dest, tmp, #:tlsgd_lo12:imm - bl __tls_get_addr - - Global Dynamic - TLS Descriptors: - adrp dest, :tlsdesc:imm - ldr tmp, [dest, #:tlsdesc_lo12:imm] - add dest, dest, #:tlsdesc_lo12:imm - blr tmp - mrs tp, tpidr_el0 - add dest, dest, tp - - Initial Exec: - mrs tp, tpidr_el0 - adrp tmp, :gottprel:imm - ldr dest, [tmp, #:gottprel_lo12:imm] - add dest, dest, tp - - Local Exec: - mrs tp, tpidr_el0 - add t0, tp, #:tprel_hi12:imm - add t0, #:tprel_lo12_nc:imm -*/ - -static void -aarch64_load_symref_appropriately (rtx dest, rtx imm, - enum aarch64_symbol_type type) -{ - switch (type) - { - case SYMBOL_SMALL_ABSOLUTE: - { - rtx tmp_reg = dest; - if (can_create_pseudo_p ()) - { - tmp_reg = gen_reg_rtx (Pmode); - } - - emit_move_insn (tmp_reg, gen_rtx_HIGH (Pmode, imm)); - emit_insn (gen_add_losym (dest, tmp_reg, imm)); - return; - } - - case SYMBOL_SMALL_GOT: - { - rtx tmp_reg = dest; - if (can_create_pseudo_p ()) - { - tmp_reg = gen_reg_rtx (Pmode); - } - emit_move_insn (tmp_reg, gen_rtx_HIGH (Pmode, imm)); - emit_insn (gen_ldr_got_small (dest, tmp_reg, imm)); - return; - } - - case SYMBOL_SMALL_TLSGD: - { - rtx insns; - rtx result = gen_rtx_REG (Pmode, R0_REGNUM); - - start_sequence (); - emit_call_insn (gen_tlsgd_small (result, imm)); - insns = get_insns (); - end_sequence (); - - RTL_CONST_CALL_P (insns) = 1; - emit_libcall_block (insns, dest, result, imm); - return; - } - - case SYMBOL_SMALL_TLSDESC: - { - rtx x0 = gen_rtx_REG (Pmode, R0_REGNUM); - rtx tp; - - emit_insn (gen_tlsdesc_small (imm)); - tp = aarch64_load_tp (NULL); - emit_insn (gen_rtx_SET (Pmode, dest, gen_rtx_PLUS (Pmode, tp, x0))); - set_unique_reg_note (get_last_insn (), REG_EQUIV, imm); - return; - } - - case SYMBOL_SMALL_GOTTPREL: - { - rtx tmp_reg = gen_reg_rtx (Pmode); - rtx tp = aarch64_load_tp (NULL); - emit_insn (gen_tlsie_small (tmp_reg, imm)); - emit_insn (gen_rtx_SET (Pmode, dest, gen_rtx_PLUS (Pmode, tp, tmp_reg))); - set_unique_reg_note (get_last_insn (), REG_EQUIV, imm); - return; - } - - case SYMBOL_SMALL_TPREL: - { - rtx tp = aarch64_load_tp (NULL); - emit_insn (gen_tlsle_small (dest, tp, imm)); - set_unique_reg_note (get_last_insn (), REG_EQUIV, imm); - return; - } - - default: - gcc_unreachable (); - } -} - -/* Emit a move from SRC to DEST. Assume that the move expanders can - handle all moves if !can_create_pseudo_p (). The distinction is - important because, unlike emit_move_insn, the move expanders know - how to force Pmode objects into the constant pool even when the - constant pool address is not itself legitimate. */ -static rtx -aarch64_emit_move (rtx dest, rtx src) -{ - return (can_create_pseudo_p () - ? emit_move_insn (dest, src) - : emit_move_insn_1 (dest, src)); -} - -void -aarch64_split_128bit_move (rtx dst, rtx src) -{ - rtx low_dst; - - gcc_assert (GET_MODE (dst) == TImode); - - if (REG_P (dst) && REG_P (src)) - { - int src_regno = REGNO (src); - int dst_regno = REGNO (dst); - - gcc_assert (GET_MODE (src) == TImode); - - /* Handle r -> w, w -> r. */ - if (FP_REGNUM_P (dst_regno) && GP_REGNUM_P (src_regno)) - { - emit_insn (gen_aarch64_movtilow_di (dst, - gen_lowpart (word_mode, src))); - emit_insn (gen_aarch64_movtihigh_di (dst, - gen_highpart (word_mode, src))); - return; - } - else if (GP_REGNUM_P (dst_regno) && FP_REGNUM_P (src_regno)) - { - emit_insn (gen_aarch64_movdi_tilow (gen_lowpart (word_mode, dst), - src)); - emit_insn (gen_aarch64_movdi_tihigh (gen_highpart (word_mode, dst), - src)); - return; - } - /* Fall through to r -> r cases. */ - } - - low_dst = gen_lowpart (word_mode, dst); - if (REG_P (low_dst) - && reg_overlap_mentioned_p (low_dst, src)) - { - aarch64_emit_move (gen_highpart (word_mode, dst), - gen_highpart_mode (word_mode, TImode, src)); - aarch64_emit_move (low_dst, gen_lowpart (word_mode, src)); - } - else - { - aarch64_emit_move (low_dst, gen_lowpart (word_mode, src)); - aarch64_emit_move (gen_highpart (word_mode, dst), - gen_highpart_mode (word_mode, TImode, src)); - } -} - -bool -aarch64_split_128bit_move_p (rtx dst, rtx src) -{ - return (! REG_P (src) - || ! (FP_REGNUM_P (REGNO (dst)) && FP_REGNUM_P (REGNO (src)))); -} - -static rtx -aarch64_force_temporary (rtx x, rtx value) -{ - if (can_create_pseudo_p ()) - return force_reg (Pmode, value); - else - { - x = aarch64_emit_move (x, value); - return x; - } -} - - -static rtx -aarch64_add_offset (enum machine_mode mode, rtx temp, rtx reg, HOST_WIDE_INT offset) -{ - if (!aarch64_plus_immediate (GEN_INT (offset), DImode)) - { - rtx high; - /* Load the full offset into a register. This - might be improvable in the future. */ - high = GEN_INT (offset); - offset = 0; - high = aarch64_force_temporary (temp, high); - reg = aarch64_force_temporary (temp, gen_rtx_PLUS (Pmode, high, reg)); - } - return plus_constant (mode, reg, offset); -} - -void -aarch64_expand_mov_immediate (rtx dest, rtx imm) -{ - enum machine_mode mode = GET_MODE (dest); - unsigned HOST_WIDE_INT mask; - int i; - bool first; - unsigned HOST_WIDE_INT val; - bool subtargets; - rtx subtarget; - int one_match, zero_match; - - gcc_assert (mode == SImode || mode == DImode); - - /* Check on what type of symbol it is. */ - if (GET_CODE (imm) == SYMBOL_REF - || GET_CODE (imm) == LABEL_REF - || GET_CODE (imm) == CONST) - { - rtx mem, base, offset; - enum aarch64_symbol_type sty; - - /* If we have (const (plus symbol offset)), separate out the offset - before we start classifying the symbol. */ - split_const (imm, &base, &offset); - - sty = aarch64_classify_symbol (base, SYMBOL_CONTEXT_ADR); - switch (sty) - { - case SYMBOL_FORCE_TO_MEM: - if (offset != const0_rtx - && targetm.cannot_force_const_mem (mode, imm)) - { - gcc_assert(can_create_pseudo_p ()); - base = aarch64_force_temporary (dest, base); - base = aarch64_add_offset (mode, NULL, base, INTVAL (offset)); - aarch64_emit_move (dest, base); - return; - } - mem = force_const_mem (mode, imm); - gcc_assert (mem); - emit_insn (gen_rtx_SET (VOIDmode, dest, mem)); - return; - - case SYMBOL_SMALL_TLSGD: - case SYMBOL_SMALL_TLSDESC: - case SYMBOL_SMALL_GOTTPREL: - case SYMBOL_SMALL_GOT: - if (offset != const0_rtx) - { - gcc_assert(can_create_pseudo_p ()); - base = aarch64_force_temporary (dest, base); - base = aarch64_add_offset (mode, NULL, base, INTVAL (offset)); - aarch64_emit_move (dest, base); - return; - } - /* FALLTHRU */ - - case SYMBOL_SMALL_TPREL: - case SYMBOL_SMALL_ABSOLUTE: - aarch64_load_symref_appropriately (dest, imm, sty); - return; - - default: - gcc_unreachable (); - } - } - - if (CONST_INT_P (imm) && aarch64_move_imm (INTVAL (imm), mode)) - { - emit_insn (gen_rtx_SET (VOIDmode, dest, imm)); - return; - } - - if (!CONST_INT_P (imm)) - { - if (GET_CODE (imm) == HIGH) - emit_insn (gen_rtx_SET (VOIDmode, dest, imm)); - else - { - rtx mem = force_const_mem (mode, imm); - gcc_assert (mem); - emit_insn (gen_rtx_SET (VOIDmode, dest, mem)); - } - - return; - } - - if (mode == SImode) - { - /* We know we can't do this in 1 insn, and we must be able to do it - in two; so don't mess around looking for sequences that don't buy - us anything. */ - emit_insn (gen_rtx_SET (VOIDmode, dest, GEN_INT (INTVAL (imm) & 0xffff))); - emit_insn (gen_insv_immsi (dest, GEN_INT (16), - GEN_INT ((INTVAL (imm) >> 16) & 0xffff))); - return; - } - - /* Remaining cases are all for DImode. */ - - val = INTVAL (imm); - subtargets = optimize && can_create_pseudo_p (); - - one_match = 0; - zero_match = 0; - mask = 0xffff; - - for (i = 0; i < 64; i += 16, mask <<= 16) - { - if ((val & mask) == 0) - zero_match++; - else if ((val & mask) == mask) - one_match++; - } - - if (one_match == 2) - { - mask = 0xffff; - for (i = 0; i < 64; i += 16, mask <<= 16) - { - if ((val & mask) != mask) - { - emit_insn (gen_rtx_SET (VOIDmode, dest, GEN_INT (val | mask))); - emit_insn (gen_insv_immdi (dest, GEN_INT (i), - GEN_INT ((val >> i) & 0xffff))); - return; - } - } - gcc_unreachable (); - } - - if (zero_match == 2) - goto simple_sequence; - - mask = 0x0ffff0000UL; - for (i = 16; i < 64; i += 16, mask <<= 16) - { - HOST_WIDE_INT comp = mask & ~(mask - 1); - - if (aarch64_uimm12_shift (val - (val & mask))) - { - subtarget = subtargets ? gen_reg_rtx (DImode) : dest; - - emit_insn (gen_rtx_SET (VOIDmode, subtarget, GEN_INT (val & mask))); - emit_insn (gen_adddi3 (dest, subtarget, - GEN_INT (val - (val & mask)))); - return; - } - else if (aarch64_uimm12_shift (-(val - ((val + comp) & mask)))) - { - subtarget = subtargets ? gen_reg_rtx (DImode) : dest; - - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT ((val + comp) & mask))); - emit_insn (gen_adddi3 (dest, subtarget, - GEN_INT (val - ((val + comp) & mask)))); - return; - } - else if (aarch64_uimm12_shift (val - ((val - comp) | ~mask))) - { - subtarget = subtargets ? gen_reg_rtx (DImode) : dest; - - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT ((val - comp) | ~mask))); - emit_insn (gen_adddi3 (dest, subtarget, - GEN_INT (val - ((val - comp) | ~mask)))); - return; - } - else if (aarch64_uimm12_shift (-(val - (val | ~mask)))) - { - subtarget = subtargets ? gen_reg_rtx (DImode) : dest; - - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT (val | ~mask))); - emit_insn (gen_adddi3 (dest, subtarget, - GEN_INT (val - (val | ~mask)))); - return; - } - } - - /* See if we can do it by arithmetically combining two - immediates. */ - for (i = 0; i < AARCH64_NUM_BITMASKS; i++) - { - int j; - mask = 0xffff; - - if (aarch64_uimm12_shift (val - aarch64_bitmasks[i]) - || aarch64_uimm12_shift (-val + aarch64_bitmasks[i])) - { - subtarget = subtargets ? gen_reg_rtx (DImode) : dest; - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT (aarch64_bitmasks[i]))); - emit_insn (gen_adddi3 (dest, subtarget, - GEN_INT (val - aarch64_bitmasks[i]))); - return; - } - - for (j = 0; j < 64; j += 16, mask <<= 16) - { - if ((aarch64_bitmasks[i] & ~mask) == (val & ~mask)) - { - emit_insn (gen_rtx_SET (VOIDmode, dest, - GEN_INT (aarch64_bitmasks[i]))); - emit_insn (gen_insv_immdi (dest, GEN_INT (j), - GEN_INT ((val >> j) & 0xffff))); - return; - } - } - } - - /* See if we can do it by logically combining two immediates. */ - for (i = 0; i < AARCH64_NUM_BITMASKS; i++) - { - if ((aarch64_bitmasks[i] & val) == aarch64_bitmasks[i]) - { - int j; - - for (j = i + 1; j < AARCH64_NUM_BITMASKS; j++) - if (val == (aarch64_bitmasks[i] | aarch64_bitmasks[j])) - { - subtarget = subtargets ? gen_reg_rtx (mode) : dest; - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT (aarch64_bitmasks[i]))); - emit_insn (gen_iordi3 (dest, subtarget, - GEN_INT (aarch64_bitmasks[j]))); - return; - } - } - else if ((val & aarch64_bitmasks[i]) == val) - { - int j; - - for (j = i + 1; j < AARCH64_NUM_BITMASKS; j++) - if (val == (aarch64_bitmasks[j] & aarch64_bitmasks[i])) - { - - subtarget = subtargets ? gen_reg_rtx (mode) : dest; - emit_insn (gen_rtx_SET (VOIDmode, subtarget, - GEN_INT (aarch64_bitmasks[j]))); - emit_insn (gen_anddi3 (dest, subtarget, - GEN_INT (aarch64_bitmasks[i]))); - return; - } - } - } - - simple_sequence: - first = true; - mask = 0xffff; - for (i = 0; i < 64; i += 16, mask <<= 16) - { - if ((val & mask) != 0) - { - if (first) - { - emit_insn (gen_rtx_SET (VOIDmode, dest, - GEN_INT (val & mask))); - first = false; - } - else - emit_insn (gen_insv_immdi (dest, GEN_INT (i), - GEN_INT ((val >> i) & 0xffff))); - } - } -} - -static bool -aarch64_function_ok_for_sibcall (tree decl, tree exp ATTRIBUTE_UNUSED) -{ - /* Indirect calls are not currently supported. */ - if (decl == NULL) - return false; - - /* Cannot tail-call to long-calls, since these are outside of the - range of a branch instruction (we could handle this if we added - support for indirect tail-calls. */ - if (aarch64_decl_is_long_call_p (decl)) - return false; - - return true; -} - -/* Implement TARGET_PASS_BY_REFERENCE. */ - -static bool -aarch64_pass_by_reference (cumulative_args_t pcum ATTRIBUTE_UNUSED, - enum machine_mode mode, - const_tree type, - bool named ATTRIBUTE_UNUSED) -{ - HOST_WIDE_INT size; - enum machine_mode dummymode; - int nregs; - - /* GET_MODE_SIZE (BLKmode) is useless since it is 0. */ - size = (mode == BLKmode && type) - ? int_size_in_bytes (type) : (int) GET_MODE_SIZE (mode); - - if (type) - { - /* Arrays always passed by reference. */ - if (TREE_CODE (type) == ARRAY_TYPE) - return true; - /* Other aggregates based on their size. */ - if (AGGREGATE_TYPE_P (type)) - size = int_size_in_bytes (type); - } - - /* Variable sized arguments are always returned by reference. */ - if (size < 0) - return true; - - /* Can this be a candidate to be passed in fp/simd register(s)? */ - if (aarch64_vfp_is_call_or_return_candidate (mode, type, - &dummymode, &nregs, - NULL)) - return false; - - /* Arguments which are variable sized or larger than 2 registers are - passed by reference unless they are a homogenous floating point - aggregate. */ - return size > 2 * UNITS_PER_WORD; -} - -/* Return TRUE if VALTYPE is padded to its least significant bits. */ -static bool -aarch64_return_in_msb (const_tree valtype) -{ - enum machine_mode dummy_mode; - int dummy_int; - - /* Never happens in little-endian mode. */ - if (!BYTES_BIG_ENDIAN) - return false; - - /* Only composite types smaller than or equal to 16 bytes can - be potentially returned in registers. */ - if (!aarch64_composite_type_p (valtype, TYPE_MODE (valtype)) - || int_size_in_bytes (valtype) <= 0 - || int_size_in_bytes (valtype) > 16) - return false; - - /* But not a composite that is an HFA (Homogeneous Floating-point Aggregate) - or an HVA (Homogeneous Short-Vector Aggregate); such a special composite - is always passed/returned in the least significant bits of fp/simd - register(s). */ - if (aarch64_vfp_is_call_or_return_candidate (TYPE_MODE (valtype), valtype, - &dummy_mode, &dummy_int, NULL)) - return false; - - return true; -} - -/* Implement TARGET_FUNCTION_VALUE. - Define how to find the value returned by a function. */ - -static rtx -aarch64_function_value (const_tree type, const_tree func, - bool outgoing ATTRIBUTE_UNUSED) -{ - enum machine_mode mode; - int unsignedp; - int count; - enum machine_mode ag_mode; - - mode = TYPE_MODE (type); - if (INTEGRAL_TYPE_P (type)) - mode = promote_function_mode (type, mode, &unsignedp, func, 1); - - if (aarch64_return_in_msb (type)) - { - HOST_WIDE_INT size = int_size_in_bytes (type); - - if (size % UNITS_PER_WORD != 0) - { - size += UNITS_PER_WORD - size % UNITS_PER_WORD; - mode = mode_for_size (size * BITS_PER_UNIT, MODE_INT, 0); - } - } - - if (aarch64_vfp_is_call_or_return_candidate (mode, type, - &ag_mode, &count, NULL)) - { - if (!aarch64_composite_type_p (type, mode)) - { - gcc_assert (count == 1 && mode == ag_mode); - return gen_rtx_REG (mode, V0_REGNUM); - } - else - { - int i; - rtx par; - - par = gen_rtx_PARALLEL (mode, rtvec_alloc (count)); - for (i = 0; i < count; i++) - { - rtx tmp = gen_rtx_REG (ag_mode, V0_REGNUM + i); - tmp = gen_rtx_EXPR_LIST (VOIDmode, tmp, - GEN_INT (i * GET_MODE_SIZE (ag_mode))); - XVECEXP (par, 0, i) = tmp; - } - return par; - } - } - else - return gen_rtx_REG (mode, R0_REGNUM); -} - -/* Implements TARGET_FUNCTION_VALUE_REGNO_P. - Return true if REGNO is the number of a hard register in which the values - of called function may come back. */ - -static bool -aarch64_function_value_regno_p (const unsigned int regno) -{ - /* Maximum of 16 bytes can be returned in the general registers. Examples - of 16-byte return values are: 128-bit integers and 16-byte small - structures (excluding homogeneous floating-point aggregates). */ - if (regno == R0_REGNUM || regno == R1_REGNUM) - return true; - - /* Up to four fp/simd registers can return a function value, e.g. a - homogeneous floating-point aggregate having four members. */ - if (regno >= V0_REGNUM && regno < V0_REGNUM + HA_MAX_NUM_FLDS) - return !TARGET_GENERAL_REGS_ONLY; - - return false; -} - -/* Implement TARGET_RETURN_IN_MEMORY. - - If the type T of the result of a function is such that - void func (T arg) - would require that arg be passed as a value in a register (or set of - registers) according to the parameter passing rules, then the result - is returned in the same registers as would be used for such an - argument. */ - -static bool -aarch64_return_in_memory (const_tree type, const_tree fndecl ATTRIBUTE_UNUSED) -{ - HOST_WIDE_INT size; - enum machine_mode ag_mode; - int count; - - if (!AGGREGATE_TYPE_P (type) - && TREE_CODE (type) != COMPLEX_TYPE - && TREE_CODE (type) != VECTOR_TYPE) - /* Simple scalar types always returned in registers. */ - return false; - - if (aarch64_vfp_is_call_or_return_candidate (TYPE_MODE (type), - type, - &ag_mode, - &count, - NULL)) - return false; - - /* Types larger than 2 registers returned in memory. */ - size = int_size_in_bytes (type); - return (size < 0 || size > 2 * UNITS_PER_WORD); -} - -static bool -aarch64_vfp_is_call_candidate (cumulative_args_t pcum_v, enum machine_mode mode, - const_tree type, int *nregs) -{ - CUMULATIVE_ARGS *pcum = get_cumulative_args (pcum_v); - return aarch64_vfp_is_call_or_return_candidate (mode, - type, - &pcum->aapcs_vfp_rmode, - nregs, - NULL); -} - -/* Given MODE and TYPE of a function argument, return the alignment in - bits. The idea is to suppress any stronger alignment requested by - the user and opt for the natural alignment (specified in AAPCS64 \S 4.1). - This is a helper function for local use only. */ - -static unsigned int -aarch64_function_arg_alignment (enum machine_mode mode, const_tree type) -{ - unsigned int alignment; - - if (type) - { - if (!integer_zerop (TYPE_SIZE (type))) - { - if (TYPE_MODE (type) == mode) - alignment = TYPE_ALIGN (type); - else - alignment = GET_MODE_ALIGNMENT (mode); - } - else - alignment = 0; - } - else - alignment = GET_MODE_ALIGNMENT (mode); - - return alignment; -} - -/* Layout a function argument according to the AAPCS64 rules. The rule - numbers refer to the rule numbers in the AAPCS64. */ - -static void -aarch64_layout_arg (cumulative_args_t pcum_v, enum machine_mode mode, - const_tree type, - bool named ATTRIBUTE_UNUSED) -{ - CUMULATIVE_ARGS *pcum = get_cumulative_args (pcum_v); - int ncrn, nvrn, nregs; - bool allocate_ncrn, allocate_nvrn; - - /* We need to do this once per argument. */ - if (pcum->aapcs_arg_processed) - return; - - pcum->aapcs_arg_processed = true; - - allocate_ncrn = (type) ? !(FLOAT_TYPE_P (type)) : !FLOAT_MODE_P (mode); - allocate_nvrn = aarch64_vfp_is_call_candidate (pcum_v, - mode, - type, - &nregs); - - /* allocate_ncrn may be false-positive, but allocate_nvrn is quite reliable. - The following code thus handles passing by SIMD/FP registers first. */ - - nvrn = pcum->aapcs_nvrn; - - /* C1 - C5 for floating point, homogenous floating point aggregates (HFA) - and homogenous short-vector aggregates (HVA). */ - if (allocate_nvrn) - { - if (nvrn + nregs <= NUM_FP_ARG_REGS) - { - pcum->aapcs_nextnvrn = nvrn + nregs; - if (!aarch64_composite_type_p (type, mode)) - { - gcc_assert (nregs == 1); - pcum->aapcs_reg = gen_rtx_REG (mode, V0_REGNUM + nvrn); - } - else - { - rtx par; - int i; - par = gen_rtx_PARALLEL (mode, rtvec_alloc (nregs)); - for (i = 0; i < nregs; i++) - { - rtx tmp = gen_rtx_REG (pcum->aapcs_vfp_rmode, - V0_REGNUM + nvrn + i); - tmp = gen_rtx_EXPR_LIST - (VOIDmode, tmp, - GEN_INT (i * GET_MODE_SIZE (pcum->aapcs_vfp_rmode))); - XVECEXP (par, 0, i) = tmp; - } - pcum->aapcs_reg = par; - } - return; - } - else - { - /* C.3 NSRN is set to 8. */ - pcum->aapcs_nextnvrn = NUM_FP_ARG_REGS; - goto on_stack; - } - } - - ncrn = pcum->aapcs_ncrn; - nregs = ((type ? int_size_in_bytes (type) : GET_MODE_SIZE (mode)) - + UNITS_PER_WORD - 1) / UNITS_PER_WORD; - - - /* C6 - C9. though the sign and zero extension semantics are - handled elsewhere. This is the case where the argument fits - entirely general registers. */ - if (allocate_ncrn && (ncrn + nregs <= NUM_ARG_REGS)) - { - unsigned int alignment = aarch64_function_arg_alignment (mode, type); - - gcc_assert (nregs == 0 || nregs == 1 || nregs == 2); - - /* C.8 if the argument has an alignment of 16 then the NGRN is - rounded up to the next even number. */ - if (nregs == 2 && alignment == 16 * BITS_PER_UNIT && ncrn % 2) - { - ++ncrn; - gcc_assert (ncrn + nregs <= NUM_ARG_REGS); - } - /* NREGS can be 0 when e.g. an empty structure is to be passed. - A reg is still generated for it, but the caller should be smart - enough not to use it. */ - if (nregs == 0 || nregs == 1 || GET_MODE_CLASS (mode) == MODE_INT) - { - pcum->aapcs_reg = gen_rtx_REG (mode, R0_REGNUM + ncrn); - } - else - { - rtx par; - int i; - - par = gen_rtx_PARALLEL (mode, rtvec_alloc (nregs)); - for (i = 0; i < nregs; i++) - { - rtx tmp = gen_rtx_REG (word_mode, R0_REGNUM + ncrn + i); - tmp = gen_rtx_EXPR_LIST (VOIDmode, tmp, - GEN_INT (i * UNITS_PER_WORD)); - XVECEXP (par, 0, i) = tmp; - } - pcum->aapcs_reg = par; - } - - pcum->aapcs_nextncrn = ncrn + nregs; - return; - } - - /* C.11 */ - pcum->aapcs_nextncrn = NUM_ARG_REGS; - - /* The argument is passed on stack; record the needed number of words for - this argument (we can re-use NREGS) and align the total size if - necessary. */ -on_stack: - pcum->aapcs_stack_words = nregs; - if (aarch64_function_arg_alignment (mode, type) == 16 * BITS_PER_UNIT) - pcum->aapcs_stack_size = AARCH64_ROUND_UP (pcum->aapcs_stack_size, - 16 / UNITS_PER_WORD) + 1; - return; -} - -/* Implement TARGET_FUNCTION_ARG. */ - -static rtx -aarch64_function_arg (cumulative_args_t pcum_v, enum machine_mode mode, - const_tree type, bool named) -{ - CUMULATIVE_ARGS *pcum = get_cumulative_args (pcum_v); - gcc_assert (pcum->pcs_variant == ARM_PCS_AAPCS64); - - if (mode == VOIDmode) - return NULL_RTX; - - aarch64_layout_arg (pcum_v, mode, type, named); - return pcum->aapcs_reg; -} - -void -aarch64_init_cumulative_args (CUMULATIVE_ARGS *pcum, - const_tree fntype ATTRIBUTE_UNUSED, - rtx libname ATTRIBUTE_UNUSED, - const_tree fndecl ATTRIBUTE_UNUSED, - unsigned n_named ATTRIBUTE_UNUSED) -{ - pcum->aapcs_ncrn = 0; - pcum->aapcs_nvrn = 0; - pcum->aapcs_nextncrn = 0; - pcum->aapcs_nextnvrn = 0; - pcum->pcs_variant = ARM_PCS_AAPCS64; - pcum->aapcs_reg = NULL_RTX; - pcum->aapcs_arg_processed = false; - pcum->aapcs_stack_words = 0; - pcum->aapcs_stack_size = 0; - - return; -} - -static void -aarch64_function_arg_advance (cumulative_args_t pcum_v, - enum machine_mode mode, - const_tree type, - bool named) -{ - CUMULATIVE_ARGS *pcum = get_cumulative_args (pcum_v); - if (pcum->pcs_variant == ARM_PCS_AAPCS64) - { - aarch64_layout_arg (pcum_v, mode, type, named); - gcc_assert ((pcum->aapcs_reg != NULL_RTX) - != (pcum->aapcs_stack_words != 0)); - pcum->aapcs_arg_processed = false; - pcum->aapcs_ncrn = pcum->aapcs_nextncrn; - pcum->aapcs_nvrn = pcum->aapcs_nextnvrn; - pcum->aapcs_stack_size += pcum->aapcs_stack_words; - pcum->aapcs_stack_words = 0; - pcum->aapcs_reg = NULL_RTX; - } -} - -bool -aarch64_function_arg_regno_p (unsigned regno) -{ - return ((GP_REGNUM_P (regno) && regno < R0_REGNUM + NUM_ARG_REGS) - || (FP_REGNUM_P (regno) && regno < V0_REGNUM + NUM_FP_ARG_REGS)); -} - -/* Implement FUNCTION_ARG_BOUNDARY. Every parameter gets at least - PARM_BOUNDARY bits of alignment, but will be given anything up - to STACK_BOUNDARY bits if the type requires it. This makes sure - that both before and after the layout of each argument, the Next - Stacked Argument Address (NSAA) will have a minimum alignment of - 8 bytes. */ - -static unsigned int -aarch64_function_arg_boundary (enum machine_mode mode, const_tree type) -{ - unsigned int alignment = aarch64_function_arg_alignment (mode, type); - - if (alignment < PARM_BOUNDARY) - alignment = PARM_BOUNDARY; - if (alignment > STACK_BOUNDARY) - alignment = STACK_BOUNDARY; - return alignment; -} - -/* For use by FUNCTION_ARG_PADDING (MODE, TYPE). - - Return true if an argument passed on the stack should be padded upwards, - i.e. if the least-significant byte of the stack slot has useful data. - - Small aggregate types are placed in the lowest memory address. - - The related parameter passing rules are B.4, C.3, C.5 and C.14. */ - -bool -aarch64_pad_arg_upward (enum machine_mode mode, const_tree type) -{ - /* On little-endian targets, the least significant byte of every stack - argument is passed at the lowest byte address of the stack slot. */ - if (!BYTES_BIG_ENDIAN) - return true; - - /* Otherwise, integral types and floating point types are padded downward: - the least significant byte of a stack argument is passed at the highest - byte address of the stack slot. */ - if (type - ? (INTEGRAL_TYPE_P (type) || SCALAR_FLOAT_TYPE_P (type)) - : (SCALAR_INT_MODE_P (mode) || SCALAR_FLOAT_MODE_P (mode))) - return false; - - /* Everything else padded upward, i.e. data in first byte of stack slot. */ - return true; -} - -/* Similarly, for use by BLOCK_REG_PADDING (MODE, TYPE, FIRST). - - It specifies padding for the last (may also be the only) - element of a block move between registers and memory. If - assuming the block is in the memory, padding upward means that - the last element is padded after its highest significant byte, - while in downward padding, the last element is padded at the - its least significant byte side. - - Small aggregates and small complex types are always padded - upwards. - - We don't need to worry about homogeneous floating-point or - short-vector aggregates; their move is not affected by the - padding direction determined here. Regardless of endianness, - each element of such an aggregate is put in the least - significant bits of a fp/simd register. - - Return !BYTES_BIG_ENDIAN if the least significant byte of the - register has useful data, and return the opposite if the most - significant byte does. */ - -bool -aarch64_pad_reg_upward (enum machine_mode mode, const_tree type, - bool first ATTRIBUTE_UNUSED) -{ - - /* Small composite types are always padded upward. */ - if (BYTES_BIG_ENDIAN && aarch64_composite_type_p (type, mode)) - { - HOST_WIDE_INT size = (type ? int_size_in_bytes (type) - : GET_MODE_SIZE (mode)); - if (size < 2 * UNITS_PER_WORD) - return true; - } - - /* Otherwise, use the default padding. */ - return !BYTES_BIG_ENDIAN; -} - -static enum machine_mode -aarch64_libgcc_cmp_return_mode (void) -{ - return SImode; -} - -static bool -aarch64_frame_pointer_required (void) -{ - /* If the function contains dynamic stack allocations, we need to - use the frame pointer to access the static parts of the frame. */ - if (cfun->calls_alloca) - return true; - - /* We may have turned flag_omit_frame_pointer on in order to have this - function called; if we did, we also set the 'faked_omit_frame_pointer' flag - and we'll check it here. - If we really did set flag_omit_frame_pointer normally, then we return false - (no frame pointer required) in all cases. */ - - if (flag_omit_frame_pointer && !faked_omit_frame_pointer) - return false; - else if (flag_omit_leaf_frame_pointer) - return !crtl->is_leaf; - return true; -} - -/* Mark the registers that need to be saved by the callee and calculate - the size of the callee-saved registers area and frame record (both FP - and LR may be omitted). */ -static void -aarch64_layout_frame (void) -{ - HOST_WIDE_INT offset = 0; - int regno; - - if (reload_completed && cfun->machine->frame.laid_out) - return; - - cfun->machine->frame.fp_lr_offset = 0; - - /* First mark all the registers that really need to be saved... */ - for (regno = R0_REGNUM; regno <= R30_REGNUM; regno++) - cfun->machine->frame.reg_offset[regno] = -1; - - for (regno = V0_REGNUM; regno <= V31_REGNUM; regno++) - cfun->machine->frame.reg_offset[regno] = -1; - - /* ... that includes the eh data registers (if needed)... */ - if (crtl->calls_eh_return) - for (regno = 0; EH_RETURN_DATA_REGNO (regno) != INVALID_REGNUM; regno++) - cfun->machine->frame.reg_offset[EH_RETURN_DATA_REGNO (regno)] = 0; - - /* ... and any callee saved register that dataflow says is live. */ - for (regno = R0_REGNUM; regno <= R30_REGNUM; regno++) - if (df_regs_ever_live_p (regno) - && !call_used_regs[regno]) - cfun->machine->frame.reg_offset[regno] = 0; - - for (regno = V0_REGNUM; regno <= V31_REGNUM; regno++) - if (df_regs_ever_live_p (regno) - && !call_used_regs[regno]) - cfun->machine->frame.reg_offset[regno] = 0; - - if (frame_pointer_needed) - { - cfun->machine->frame.reg_offset[R30_REGNUM] = 0; - cfun->machine->frame.reg_offset[R29_REGNUM] = 0; - cfun->machine->frame.hardfp_offset = 2 * UNITS_PER_WORD; - } - - /* Now assign stack slots for them. */ - for (regno = R0_REGNUM; regno <= R28_REGNUM; regno++) - if (cfun->machine->frame.reg_offset[regno] != -1) - { - cfun->machine->frame.reg_offset[regno] = offset; - offset += UNITS_PER_WORD; - } - - for (regno = V0_REGNUM; regno <= V31_REGNUM; regno++) - if (cfun->machine->frame.reg_offset[regno] != -1) - { - cfun->machine->frame.reg_offset[regno] = offset; - offset += UNITS_PER_WORD; - } - - if (frame_pointer_needed) - { - cfun->machine->frame.reg_offset[R29_REGNUM] = offset; - offset += UNITS_PER_WORD; - cfun->machine->frame.fp_lr_offset = UNITS_PER_WORD; - } - - if (cfun->machine->frame.reg_offset[R30_REGNUM] != -1) - { - cfun->machine->frame.reg_offset[R30_REGNUM] = offset; - offset += UNITS_PER_WORD; - cfun->machine->frame.fp_lr_offset += UNITS_PER_WORD; - } - - cfun->machine->frame.padding0 = - (AARCH64_ROUND_UP (offset, STACK_BOUNDARY / BITS_PER_UNIT) - offset); - offset = AARCH64_ROUND_UP (offset, STACK_BOUNDARY / BITS_PER_UNIT); - - cfun->machine->frame.saved_regs_size = offset; - cfun->machine->frame.laid_out = true; -} - -/* Make the last instruction frame-related and note that it performs - the operation described by FRAME_PATTERN. */ - -static void -aarch64_set_frame_expr (rtx frame_pattern) -{ - rtx insn; - - insn = get_last_insn (); - RTX_FRAME_RELATED_P (insn) = 1; - RTX_FRAME_RELATED_P (frame_pattern) = 1; - REG_NOTES (insn) = alloc_EXPR_LIST (REG_FRAME_RELATED_EXPR, - frame_pattern, - REG_NOTES (insn)); -} - -static bool -aarch64_register_saved_on_entry (int regno) -{ - return cfun->machine->frame.reg_offset[regno] != -1; -} - - -static void -aarch64_save_or_restore_fprs (int start_offset, int increment, - bool restore, rtx base_rtx) - -{ - unsigned regno; - unsigned regno2; - rtx insn; - rtx (*gen_mem_ref)(enum machine_mode, rtx) = (frame_pointer_needed)? gen_frame_mem : gen_rtx_MEM; - - - for (regno = V0_REGNUM; regno <= V31_REGNUM; regno++) - { - if (aarch64_register_saved_on_entry (regno)) - { - rtx mem; - mem = gen_mem_ref (DFmode, - plus_constant (Pmode, - base_rtx, - start_offset)); - - for (regno2 = regno + 1; - regno2 <= V31_REGNUM - && !aarch64_register_saved_on_entry (regno2); - regno2++) - { - /* Empty loop. */ - } - if (regno2 <= V31_REGNUM && - aarch64_register_saved_on_entry (regno2)) - { - rtx mem2; - /* Next highest register to be saved. */ - mem2 = gen_mem_ref (DFmode, - plus_constant - (Pmode, - base_rtx, - start_offset + increment)); - if (restore == false) - { - insn = emit_insn - ( gen_store_pairdf (mem, gen_rtx_REG (DFmode, regno), - mem2, gen_rtx_REG (DFmode, regno2))); - - } - else - { - insn = emit_insn - ( gen_load_pairdf (gen_rtx_REG (DFmode, regno), mem, - gen_rtx_REG (DFmode, regno2), mem2)); - - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DFmode, regno)); - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DFmode, regno2)); - } - - /* The first part of a frame-related parallel insn - is always assumed to be relevant to the frame - calculations; subsequent parts, are only - frame-related if explicitly marked. */ - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, - 1)) = 1; - regno = regno2; - start_offset += increment * 2; - } - else - { - if (restore == false) - insn = emit_move_insn (mem, gen_rtx_REG (DFmode, regno)); - else - { - insn = emit_move_insn (gen_rtx_REG (DFmode, regno), mem); - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DImode, regno)); - } - start_offset += increment; - } - RTX_FRAME_RELATED_P (insn) = 1; - } - } - -} - - -/* offset from the stack pointer of where the saves and - restore's have to happen. */ -static void -aarch64_save_or_restore_callee_save_registers (HOST_WIDE_INT offset, - bool restore) -{ - rtx insn; - rtx base_rtx = stack_pointer_rtx; - HOST_WIDE_INT start_offset = offset; - HOST_WIDE_INT increment = UNITS_PER_WORD; - rtx (*gen_mem_ref)(enum machine_mode, rtx) = (frame_pointer_needed)? gen_frame_mem : gen_rtx_MEM; - unsigned limit = (frame_pointer_needed)? R28_REGNUM: R30_REGNUM; - unsigned regno; - unsigned regno2; - - for (regno = R0_REGNUM; regno <= limit; regno++) - { - if (aarch64_register_saved_on_entry (regno)) - { - rtx mem; - mem = gen_mem_ref (Pmode, - plus_constant (Pmode, - base_rtx, - start_offset)); - - for (regno2 = regno + 1; - regno2 <= limit - && !aarch64_register_saved_on_entry (regno2); - regno2++) - { - /* Empty loop. */ - } - if (regno2 <= limit && - aarch64_register_saved_on_entry (regno2)) - { - rtx mem2; - /* Next highest register to be saved. */ - mem2 = gen_mem_ref (Pmode, - plus_constant - (Pmode, - base_rtx, - start_offset + increment)); - if (restore == false) - { - insn = emit_insn - ( gen_store_pairdi (mem, gen_rtx_REG (DImode, regno), - mem2, gen_rtx_REG (DImode, regno2))); - - } - else - { - insn = emit_insn - ( gen_load_pairdi (gen_rtx_REG (DImode, regno), mem, - gen_rtx_REG (DImode, regno2), mem2)); - - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DImode, regno)); - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DImode, regno2)); - } - - /* The first part of a frame-related parallel insn - is always assumed to be relevant to the frame - calculations; subsequent parts, are only - frame-related if explicitly marked. */ - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, - 1)) = 1; - regno = regno2; - start_offset += increment * 2; - } - else - { - if (restore == false) - insn = emit_move_insn (mem, gen_rtx_REG (DImode, regno)); - else - { - insn = emit_move_insn (gen_rtx_REG (DImode, regno), mem); - add_reg_note (insn, REG_CFA_RESTORE, gen_rtx_REG (DImode, regno)); - } - start_offset += increment; - } - RTX_FRAME_RELATED_P (insn) = 1; - } - } - - aarch64_save_or_restore_fprs (start_offset, increment, restore, base_rtx); - -} - -/* AArch64 stack frames generated by this compiler look like: - - +-------------------------------+ - | | - | incoming stack arguments | - | | - +-------------------------------+ <-- arg_pointer_rtx - | | - | callee-allocated save area | - | for register varargs | - | | - +-------------------------------+ - | | - | local variables | - | | - +-------------------------------+ <-- frame_pointer_rtx - | | - | callee-saved registers | - | | - +-------------------------------+ - | LR' | - +-------------------------------+ - | FP' | - P +-------------------------------+ <-- hard_frame_pointer_rtx - | dynamic allocation | - +-------------------------------+ - | | - | outgoing stack arguments | - | | - +-------------------------------+ <-- stack_pointer_rtx - - Dynamic stack allocations such as alloca insert data at point P. - They decrease stack_pointer_rtx but leave frame_pointer_rtx and - hard_frame_pointer_rtx unchanged. */ - -/* Generate the prologue instructions for entry into a function. - Establish the stack frame by decreasing the stack pointer with a - properly calculated size and, if necessary, create a frame record - filled with the values of LR and previous frame pointer. The - current FP is also set up is it is in use. */ - -void -aarch64_expand_prologue (void) -{ - /* sub sp, sp, #<frame_size> - stp {fp, lr}, [sp, #<frame_size> - 16] - add fp, sp, #<frame_size> - hardfp_offset - stp {cs_reg}, [fp, #-16] etc. - - sub sp, sp, <final_adjustment_if_any> - */ - HOST_WIDE_INT original_frame_size; /* local variables + vararg save */ - HOST_WIDE_INT frame_size, offset; - HOST_WIDE_INT fp_offset; /* FP offset from SP */ - rtx insn; - - aarch64_layout_frame (); - original_frame_size = get_frame_size () + cfun->machine->saved_varargs_size; - gcc_assert ((!cfun->machine->saved_varargs_size || cfun->stdarg) - && (cfun->stdarg || !cfun->machine->saved_varargs_size)); - frame_size = (original_frame_size + cfun->machine->frame.saved_regs_size - + crtl->outgoing_args_size); - offset = frame_size = AARCH64_ROUND_UP (frame_size, - STACK_BOUNDARY / BITS_PER_UNIT); - - if (flag_stack_usage_info) - current_function_static_stack_size = frame_size; - - fp_offset = (offset - - original_frame_size - - cfun->machine->frame.saved_regs_size); - - /* Store pairs and load pairs have a range only -512 to 504. */ - if (offset >= 512) - { - /* When the frame has a large size, an initial decrease is done on - the stack pointer to jump over the callee-allocated save area for - register varargs, the local variable area and/or the callee-saved - register area. This will allow the pre-index write-back - store pair instructions to be used for setting up the stack frame - efficiently. */ - offset = original_frame_size + cfun->machine->frame.saved_regs_size; - if (offset >= 512) - offset = cfun->machine->frame.saved_regs_size; - - frame_size -= (offset + crtl->outgoing_args_size); - fp_offset = 0; - - if (frame_size >= 0x1000000) - { - rtx op0 = gen_rtx_REG (Pmode, IP0_REGNUM); - emit_move_insn (op0, GEN_INT (-frame_size)); - emit_insn (gen_add2_insn (stack_pointer_rtx, op0)); - aarch64_set_frame_expr (gen_rtx_SET - (Pmode, stack_pointer_rtx, - gen_rtx_PLUS (Pmode, - stack_pointer_rtx, - GEN_INT (-frame_size)))); - } - else if (frame_size > 0) - { - if ((frame_size & 0xfff) != frame_size) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT (-(frame_size - & ~(HOST_WIDE_INT)0xfff)))); - RTX_FRAME_RELATED_P (insn) = 1; - } - if ((frame_size & 0xfff) != 0) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT (-(frame_size - & (HOST_WIDE_INT)0xfff)))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } - } - else - frame_size = -1; - - if (offset > 0) - { - /* Save the frame pointer and lr if the frame pointer is needed - first. Make the frame pointer point to the location of the - old frame pointer on the stack. */ - if (frame_pointer_needed) - { - rtx mem_fp, mem_lr; - - if (fp_offset) - { - insn = emit_insn (gen_add2_insn (stack_pointer_rtx, - GEN_INT (-offset))); - RTX_FRAME_RELATED_P (insn) = 1; - aarch64_set_frame_expr (gen_rtx_SET - (Pmode, stack_pointer_rtx, - gen_rtx_MINUS (Pmode, - stack_pointer_rtx, - GEN_INT (offset)))); - mem_fp = gen_frame_mem (DImode, - plus_constant (Pmode, - stack_pointer_rtx, - fp_offset)); - mem_lr = gen_frame_mem (DImode, - plus_constant (Pmode, - stack_pointer_rtx, - fp_offset - + UNITS_PER_WORD)); - insn = emit_insn (gen_store_pairdi (mem_fp, - hard_frame_pointer_rtx, - mem_lr, - gen_rtx_REG (DImode, - LR_REGNUM))); - } - else - { - insn = emit_insn (gen_storewb_pairdi_di - (stack_pointer_rtx, stack_pointer_rtx, - hard_frame_pointer_rtx, - gen_rtx_REG (DImode, LR_REGNUM), - GEN_INT (-offset), - GEN_INT (GET_MODE_SIZE (DImode) - offset))); - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, 2)) = 1; - } - - /* The first part of a frame-related parallel insn is always - assumed to be relevant to the frame calculations; - subsequent parts, are only frame-related if explicitly - marked. */ - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, 1)) = 1; - RTX_FRAME_RELATED_P (insn) = 1; - - /* Set up frame pointer to point to the location of the - previous frame pointer on the stack. */ - insn = emit_insn (gen_add3_insn (hard_frame_pointer_rtx, - stack_pointer_rtx, - GEN_INT (fp_offset))); - aarch64_set_frame_expr (gen_rtx_SET - (Pmode, hard_frame_pointer_rtx, - gen_rtx_PLUS (Pmode, - stack_pointer_rtx, - GEN_INT (fp_offset)))); - RTX_FRAME_RELATED_P (insn) = 1; - insn = emit_insn (gen_stack_tie (stack_pointer_rtx, - hard_frame_pointer_rtx)); - } - else - { - insn = emit_insn (gen_add2_insn (stack_pointer_rtx, - GEN_INT (-offset))); - RTX_FRAME_RELATED_P (insn) = 1; - } - - aarch64_save_or_restore_callee_save_registers - (fp_offset + cfun->machine->frame.hardfp_offset, 0); - } - - /* when offset >= 512, - sub sp, sp, #<outgoing_args_size> */ - if (frame_size > -1) - { - if (crtl->outgoing_args_size > 0) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT (- crtl->outgoing_args_size))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } -} - -/* Generate the epilogue instructions for returning from a function. */ -void -aarch64_expand_epilogue (bool for_sibcall) -{ - HOST_WIDE_INT original_frame_size, frame_size, offset; - HOST_WIDE_INT fp_offset; - rtx insn; - rtx cfa_reg; - - aarch64_layout_frame (); - original_frame_size = get_frame_size () + cfun->machine->saved_varargs_size; - frame_size = (original_frame_size + cfun->machine->frame.saved_regs_size - + crtl->outgoing_args_size); - offset = frame_size = AARCH64_ROUND_UP (frame_size, - STACK_BOUNDARY / BITS_PER_UNIT); - - fp_offset = (offset - - original_frame_size - - cfun->machine->frame.saved_regs_size); - - cfa_reg = frame_pointer_needed ? hard_frame_pointer_rtx : stack_pointer_rtx; - - /* Store pairs and load pairs have a range only -512 to 504. */ - if (offset >= 512) - { - offset = original_frame_size + cfun->machine->frame.saved_regs_size; - if (offset >= 512) - offset = cfun->machine->frame.saved_regs_size; - - frame_size -= (offset + crtl->outgoing_args_size); - fp_offset = 0; - if (!frame_pointer_needed && crtl->outgoing_args_size > 0) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT (crtl->outgoing_args_size))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } - else - frame_size = -1; - - /* If there were outgoing arguments or we've done dynamic stack - allocation, then restore the stack pointer from the frame - pointer. This is at most one insn and more efficient than using - GCC's internal mechanism. */ - if (frame_pointer_needed - && (crtl->outgoing_args_size || cfun->calls_alloca)) - { - insn = emit_insn (gen_add3_insn (stack_pointer_rtx, - hard_frame_pointer_rtx, - GEN_INT (- fp_offset))); - RTX_FRAME_RELATED_P (insn) = 1; - /* As SP is set to (FP - fp_offset), according to the rules in - dwarf2cfi.c:dwarf2out_frame_debug_expr, CFA should be calculated - from the value of SP from now on. */ - cfa_reg = stack_pointer_rtx; - } - - aarch64_save_or_restore_callee_save_registers - (fp_offset + cfun->machine->frame.hardfp_offset, 1); - - /* Restore the frame pointer and lr if the frame pointer is needed. */ - if (offset > 0) - { - if (frame_pointer_needed) - { - rtx mem_fp, mem_lr; - - if (fp_offset) - { - mem_fp = gen_frame_mem (DImode, - plus_constant (Pmode, - stack_pointer_rtx, - fp_offset)); - mem_lr = gen_frame_mem (DImode, - plus_constant (Pmode, - stack_pointer_rtx, - fp_offset - + UNITS_PER_WORD)); - insn = emit_insn (gen_load_pairdi (hard_frame_pointer_rtx, - mem_fp, - gen_rtx_REG (DImode, - LR_REGNUM), - mem_lr)); - } - else - { - insn = emit_insn (gen_loadwb_pairdi_di - (stack_pointer_rtx, - stack_pointer_rtx, - hard_frame_pointer_rtx, - gen_rtx_REG (DImode, LR_REGNUM), - GEN_INT (offset), - GEN_INT (GET_MODE_SIZE (DImode) + offset))); - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, 2)) = 1; - add_reg_note (insn, REG_CFA_ADJUST_CFA, - (gen_rtx_SET (Pmode, stack_pointer_rtx, - plus_constant (Pmode, cfa_reg, - offset)))); - } - - /* The first part of a frame-related parallel insn - is always assumed to be relevant to the frame - calculations; subsequent parts, are only - frame-related if explicitly marked. */ - RTX_FRAME_RELATED_P (XVECEXP (PATTERN (insn), 0, 1)) = 1; - RTX_FRAME_RELATED_P (insn) = 1; - add_reg_note (insn, REG_CFA_RESTORE, hard_frame_pointer_rtx); - add_reg_note (insn, REG_CFA_RESTORE, - gen_rtx_REG (DImode, LR_REGNUM)); - - if (fp_offset) - { - insn = emit_insn (gen_add2_insn (stack_pointer_rtx, - GEN_INT (offset))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } - else - { - insn = emit_insn (gen_add2_insn (stack_pointer_rtx, - GEN_INT (offset))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } - - /* Stack adjustment for exception handler. */ - if (crtl->calls_eh_return) - { - /* We need to unwind the stack by the offset computed by - EH_RETURN_STACKADJ_RTX. However, at this point the CFA is - based on SP. Ideally we would update the SP and define the - CFA along the lines of: - - SP = SP + EH_RETURN_STACKADJ_RTX - (regnote CFA = SP - EH_RETURN_STACKADJ_RTX) - - However the dwarf emitter only understands a constant - register offset. - - The solution choosen here is to use the otherwise unused IP0 - as a temporary register to hold the current SP value. The - CFA is described using IP0 then SP is modified. */ - - rtx ip0 = gen_rtx_REG (DImode, IP0_REGNUM); - - insn = emit_move_insn (ip0, stack_pointer_rtx); - add_reg_note (insn, REG_CFA_DEF_CFA, ip0); - RTX_FRAME_RELATED_P (insn) = 1; - - emit_insn (gen_add2_insn (stack_pointer_rtx, EH_RETURN_STACKADJ_RTX)); - - /* Ensure the assignment to IP0 does not get optimized away. */ - emit_use (ip0); - } - - if (frame_size > -1) - { - if (frame_size >= 0x1000000) - { - rtx op0 = gen_rtx_REG (Pmode, IP0_REGNUM); - emit_move_insn (op0, GEN_INT (frame_size)); - emit_insn (gen_add2_insn (stack_pointer_rtx, op0)); - aarch64_set_frame_expr (gen_rtx_SET - (Pmode, stack_pointer_rtx, - gen_rtx_PLUS (Pmode, - stack_pointer_rtx, - GEN_INT (frame_size)))); - } - else if (frame_size > 0) - { - if ((frame_size & 0xfff) != 0) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT ((frame_size - & (HOST_WIDE_INT) 0xfff)))); - RTX_FRAME_RELATED_P (insn) = 1; - } - if ((frame_size & 0xfff) != frame_size) - { - insn = emit_insn (gen_add2_insn - (stack_pointer_rtx, - GEN_INT ((frame_size - & ~ (HOST_WIDE_INT) 0xfff)))); - RTX_FRAME_RELATED_P (insn) = 1; - } - } - - aarch64_set_frame_expr (gen_rtx_SET (Pmode, stack_pointer_rtx, - gen_rtx_PLUS (Pmode, - stack_pointer_rtx, - GEN_INT (offset)))); - } - - emit_use (gen_rtx_REG (DImode, LR_REGNUM)); - if (!for_sibcall) - emit_jump_insn (ret_rtx); -} - -/* Return the place to copy the exception unwinding return address to. - This will probably be a stack slot, but could (in theory be the - return register). */ -rtx -aarch64_final_eh_return_addr (void) -{ - HOST_WIDE_INT original_frame_size, frame_size, offset, fp_offset; - aarch64_layout_frame (); - original_frame_size = get_frame_size () + cfun->machine->saved_varargs_size; - frame_size = (original_frame_size + cfun->machine->frame.saved_regs_size - + crtl->outgoing_args_size); - offset = frame_size = AARCH64_ROUND_UP (frame_size, - STACK_BOUNDARY / BITS_PER_UNIT); - fp_offset = offset - - original_frame_size - - cfun->machine->frame.saved_regs_size; - - if (cfun->machine->frame.reg_offset[LR_REGNUM] < 0) - return gen_rtx_REG (DImode, LR_REGNUM); - - /* DSE and CSELIB do not detect an alias between sp+k1 and fp+k2. This can - result in a store to save LR introduced by builtin_eh_return () being - incorrectly deleted because the alias is not detected. - So in the calculation of the address to copy the exception unwinding - return address to, we note 2 cases. - If FP is needed and the fp_offset is 0, it means that SP = FP and hence - we return a SP-relative location since all the addresses are SP-relative - in this case. This prevents the store from being optimized away. - If the fp_offset is not 0, then the addresses will be FP-relative and - therefore we return a FP-relative location. */ - - if (frame_pointer_needed) - { - if (fp_offset) - return gen_frame_mem (DImode, - plus_constant (Pmode, hard_frame_pointer_rtx, UNITS_PER_WORD)); - else - return gen_frame_mem (DImode, - plus_constant (Pmode, stack_pointer_rtx, UNITS_PER_WORD)); - } - - /* If FP is not needed, we calculate the location of LR, which would be - at the top of the saved registers block. */ - - return gen_frame_mem (DImode, - plus_constant (Pmode, - stack_pointer_rtx, - fp_offset - + cfun->machine->frame.saved_regs_size - - 2 * UNITS_PER_WORD)); -} - -/* Output code to build up a constant in a register. */ -static void -aarch64_build_constant (int regnum, HOST_WIDE_INT val) -{ - if (aarch64_bitmask_imm (val, DImode)) - emit_move_insn (gen_rtx_REG (Pmode, regnum), GEN_INT (val)); - else - { - int i; - int ncount = 0; - int zcount = 0; - HOST_WIDE_INT valp = val >> 16; - HOST_WIDE_INT valm; - HOST_WIDE_INT tval; - - for (i = 16; i < 64; i += 16) - { - valm = (valp & 0xffff); - - if (valm != 0) - ++ zcount; - - if (valm != 0xffff) - ++ ncount; - - valp >>= 16; - } - - /* zcount contains the number of additional MOVK instructions - required if the constant is built up with an initial MOVZ instruction, - while ncount is the number of MOVK instructions required if starting - with a MOVN instruction. Choose the sequence that yields the fewest - number of instructions, preferring MOVZ instructions when they are both - the same. */ - if (ncount < zcount) - { - emit_move_insn (gen_rtx_REG (Pmode, regnum), - GEN_INT ((~val) & 0xffff)); - tval = 0xffff; - } - else - { - emit_move_insn (gen_rtx_REG (Pmode, regnum), - GEN_INT (val & 0xffff)); - tval = 0; - } - - val >>= 16; - - for (i = 16; i < 64; i += 16) - { - if ((val & 0xffff) != tval) - emit_insn (gen_insv_immdi (gen_rtx_REG (Pmode, regnum), - GEN_INT (i), GEN_INT (val & 0xffff))); - val >>= 16; - } - } -} - -static void -aarch64_add_constant (int regnum, int scratchreg, HOST_WIDE_INT delta) -{ - HOST_WIDE_INT mdelta = delta; - rtx this_rtx = gen_rtx_REG (Pmode, regnum); - rtx scratch_rtx = gen_rtx_REG (Pmode, scratchreg); - - if (mdelta < 0) - mdelta = -mdelta; - - if (mdelta >= 4096 * 4096) - { - aarch64_build_constant (scratchreg, delta); - emit_insn (gen_add3_insn (this_rtx, this_rtx, scratch_rtx)); - } - else if (mdelta > 0) - { - if (mdelta >= 4096) - { - emit_insn (gen_rtx_SET (Pmode, scratch_rtx, GEN_INT (mdelta / 4096))); - rtx shift = gen_rtx_ASHIFT (Pmode, scratch_rtx, GEN_INT (12)); - if (delta < 0) - emit_insn (gen_rtx_SET (Pmode, this_rtx, - gen_rtx_MINUS (Pmode, this_rtx, shift))); - else - emit_insn (gen_rtx_SET (Pmode, this_rtx, - gen_rtx_PLUS (Pmode, this_rtx, shift))); - } - if (mdelta % 4096 != 0) - { - scratch_rtx = GEN_INT ((delta < 0 ? -1 : 1) * (mdelta % 4096)); - emit_insn (gen_rtx_SET (Pmode, this_rtx, - gen_rtx_PLUS (Pmode, this_rtx, scratch_rtx))); - } - } -} - -/* Output code to add DELTA to the first argument, and then jump - to FUNCTION. Used for C++ multiple inheritance. */ -static void -aarch64_output_mi_thunk (FILE *file, tree thunk ATTRIBUTE_UNUSED, - HOST_WIDE_INT delta, - HOST_WIDE_INT vcall_offset, - tree function) -{ - /* The this pointer is always in x0. Note that this differs from - Arm where the this pointer maybe bumped to r1 if r0 is required - to return a pointer to an aggregate. On AArch64 a result value - pointer will be in x8. */ - int this_regno = R0_REGNUM; - rtx this_rtx, temp0, temp1, addr, insn, funexp; - - reload_completed = 1; - emit_note (NOTE_INSN_PROLOGUE_END); - - if (vcall_offset == 0) - aarch64_add_constant (this_regno, IP1_REGNUM, delta); - else - { - gcc_assert ((vcall_offset & 0x7) == 0); - - this_rtx = gen_rtx_REG (Pmode, this_regno); - temp0 = gen_rtx_REG (Pmode, IP0_REGNUM); - temp1 = gen_rtx_REG (Pmode, IP1_REGNUM); - - addr = this_rtx; - if (delta != 0) - { - if (delta >= -256 && delta < 256) - addr = gen_rtx_PRE_MODIFY (Pmode, this_rtx, - plus_constant (Pmode, this_rtx, delta)); - else - aarch64_add_constant (this_regno, IP1_REGNUM, delta); - } - - aarch64_emit_move (temp0, gen_rtx_MEM (Pmode, addr)); - - if (vcall_offset >= -256 && vcall_offset < 32768) - addr = plus_constant (Pmode, temp0, vcall_offset); - else - { - aarch64_build_constant (IP1_REGNUM, vcall_offset); - addr = gen_rtx_PLUS (Pmode, temp0, temp1); - } - - aarch64_emit_move (temp1, gen_rtx_MEM (Pmode,addr)); - emit_insn (gen_add2_insn (this_rtx, temp1)); - } - - /* Generate a tail call to the target function. */ - if (!TREE_USED (function)) - { - assemble_external (function); - TREE_USED (function) = 1; - } - funexp = XEXP (DECL_RTL (function), 0); - funexp = gen_rtx_MEM (FUNCTION_MODE, funexp); - insn = emit_call_insn (gen_sibcall (funexp, const0_rtx, NULL_RTX)); - SIBLING_CALL_P (insn) = 1; - - insn = get_insns (); - shorten_branches (insn); - final_start_function (insn, file, 1); - final (insn, file, 1); - final_end_function (); - - /* Stop pretending to be a post-reload pass. */ - reload_completed = 0; -} - -static int -aarch64_tls_operand_p_1 (rtx *x, void *data ATTRIBUTE_UNUSED) -{ - if (GET_CODE (*x) == SYMBOL_REF) - return SYMBOL_REF_TLS_MODEL (*x) != 0; - - /* Don't recurse into UNSPEC_TLS looking for TLS symbols; these are - TLS offsets, not real symbol references. */ - if (GET_CODE (*x) == UNSPEC - && XINT (*x, 1) == UNSPEC_TLS) - return -1; - - return 0; -} - -static bool -aarch64_tls_referenced_p (rtx x) -{ - if (!TARGET_HAVE_TLS) - return false; - - return for_each_rtx (&x, aarch64_tls_operand_p_1, NULL); -} - - -static int -aarch64_bitmasks_cmp (const void *i1, const void *i2) -{ - const unsigned HOST_WIDE_INT *imm1 = (const unsigned HOST_WIDE_INT *) i1; - const unsigned HOST_WIDE_INT *imm2 = (const unsigned HOST_WIDE_INT *) i2; - - if (*imm1 < *imm2) - return -1; - if (*imm1 > *imm2) - return +1; - return 0; -} - - -static void -aarch64_build_bitmask_table (void) -{ - unsigned HOST_WIDE_INT mask, imm; - unsigned int log_e, e, s, r; - unsigned int nimms = 0; - - for (log_e = 1; log_e <= 6; log_e++) - { - e = 1 << log_e; - if (e == 64) - mask = ~(HOST_WIDE_INT) 0; - else - mask = ((HOST_WIDE_INT) 1 << e) - 1; - for (s = 1; s < e; s++) - { - for (r = 0; r < e; r++) - { - /* set s consecutive bits to 1 (s < 64) */ - imm = ((unsigned HOST_WIDE_INT)1 << s) - 1; - /* rotate right by r */ - if (r != 0) - imm = ((imm >> r) | (imm << (e - r))) & mask; - /* replicate the constant depending on SIMD size */ - switch (log_e) { - case 1: imm |= (imm << 2); - case 2: imm |= (imm << 4); - case 3: imm |= (imm << 8); - case 4: imm |= (imm << 16); - case 5: imm |= (imm << 32); - case 6: - break; - default: - gcc_unreachable (); - } - gcc_assert (nimms < AARCH64_NUM_BITMASKS); - aarch64_bitmasks[nimms++] = imm; - } - } - } - - gcc_assert (nimms == AARCH64_NUM_BITMASKS); - qsort (aarch64_bitmasks, nimms, sizeof (aarch64_bitmasks[0]), - aarch64_bitmasks_cmp); -} - - -/* Return true if val can be encoded as a 12-bit unsigned immediate with - a left shift of 0 or 12 bits. */ -bool -aarch64_uimm12_shift (HOST_WIDE_INT val) -{ - return ((val & (((HOST_WIDE_INT) 0xfff) << 0)) == val - || (val & (((HOST_WIDE_INT) 0xfff) << 12)) == val - ); -} - - -/* Return true if val is an immediate that can be loaded into a - register by a MOVZ instruction. */ -static bool -aarch64_movw_imm (HOST_WIDE_INT val, enum machine_mode mode) -{ - if (GET_MODE_SIZE (mode) > 4) - { - if ((val & (((HOST_WIDE_INT) 0xffff) << 32)) == val - || (val & (((HOST_WIDE_INT) 0xffff) << 48)) == val) - return 1; - } - else - { - /* Ignore sign extension. */ - val &= (HOST_WIDE_INT) 0xffffffff; - } - return ((val & (((HOST_WIDE_INT) 0xffff) << 0)) == val - || (val & (((HOST_WIDE_INT) 0xffff) << 16)) == val); -} - - -/* Return true if val is a valid bitmask immediate. */ -bool -aarch64_bitmask_imm (HOST_WIDE_INT val, enum machine_mode mode) -{ - if (GET_MODE_SIZE (mode) < 8) - { - /* Replicate bit pattern. */ - val &= (HOST_WIDE_INT) 0xffffffff; - val |= val << 32; - } - return bsearch (&val, aarch64_bitmasks, AARCH64_NUM_BITMASKS, - sizeof (aarch64_bitmasks[0]), aarch64_bitmasks_cmp) != NULL; -} - - -/* Return true if val is an immediate that can be loaded into a - register in a single instruction. */ -bool -aarch64_move_imm (HOST_WIDE_INT val, enum machine_mode mode) -{ - if (aarch64_movw_imm (val, mode) || aarch64_movw_imm (~val, mode)) - return 1; - return aarch64_bitmask_imm (val, mode); -} - -static bool -aarch64_cannot_force_const_mem (enum machine_mode mode ATTRIBUTE_UNUSED, rtx x) -{ - rtx base, offset; - if (GET_CODE (x) == HIGH) - return true; - - split_const (x, &base, &offset); - if (GET_CODE (base) == SYMBOL_REF || GET_CODE (base) == LABEL_REF) - return (aarch64_classify_symbol (base, SYMBOL_CONTEXT_ADR) != SYMBOL_FORCE_TO_MEM); - - return aarch64_tls_referenced_p (x); -} - -/* Return true if register REGNO is a valid index register. - STRICT_P is true if REG_OK_STRICT is in effect. */ - -bool -aarch64_regno_ok_for_index_p (int regno, bool strict_p) -{ - if (!HARD_REGISTER_NUM_P (regno)) - { - if (!strict_p) - return true; - - if (!reg_renumber) - return false; - - regno = reg_renumber[regno]; - } - return GP_REGNUM_P (regno); -} - -/* Return true if register REGNO is a valid base register for mode MODE. - STRICT_P is true if REG_OK_STRICT is in effect. */ - -bool -aarch64_regno_ok_for_base_p (int regno, bool strict_p) -{ - if (!HARD_REGISTER_NUM_P (regno)) - { - if (!strict_p) - return true; - - if (!reg_renumber) - return false; - - regno = reg_renumber[regno]; - } - - /* The fake registers will be eliminated to either the stack or - hard frame pointer, both of which are usually valid base registers. - Reload deals with the cases where the eliminated form isn't valid. */ - return (GP_REGNUM_P (regno) - || regno == SP_REGNUM - || regno == FRAME_POINTER_REGNUM - || regno == ARG_POINTER_REGNUM); -} - -/* Return true if X is a valid base register for mode MODE. - STRICT_P is true if REG_OK_STRICT is in effect. */ - -static bool -aarch64_base_register_rtx_p (rtx x, bool strict_p) -{ - if (!strict_p && GET_CODE (x) == SUBREG) - x = SUBREG_REG (x); - - return (REG_P (x) && aarch64_regno_ok_for_base_p (REGNO (x), strict_p)); -} - -/* Return true if address offset is a valid index. If it is, fill in INFO - appropriately. STRICT_P is true if REG_OK_STRICT is in effect. */ - -static bool -aarch64_classify_index (struct aarch64_address_info *info, rtx x, - enum machine_mode mode, bool strict_p) -{ - enum aarch64_address_type type; - rtx index; - int shift; - - /* (reg:P) */ - if ((REG_P (x) || GET_CODE (x) == SUBREG) - && GET_MODE (x) == Pmode) - { - type = ADDRESS_REG_REG; - index = x; - shift = 0; - } - /* (sign_extend:DI (reg:SI)) */ - else if ((GET_CODE (x) == SIGN_EXTEND - || GET_CODE (x) == ZERO_EXTEND) - && GET_MODE (x) == DImode - && GET_MODE (XEXP (x, 0)) == SImode) - { - type = (GET_CODE (x) == SIGN_EXTEND) - ? ADDRESS_REG_SXTW : ADDRESS_REG_UXTW; - index = XEXP (x, 0); - shift = 0; - } - /* (mult:DI (sign_extend:DI (reg:SI)) (const_int scale)) */ - else if (GET_CODE (x) == MULT - && (GET_CODE (XEXP (x, 0)) == SIGN_EXTEND - || GET_CODE (XEXP (x, 0)) == ZERO_EXTEND) - && GET_MODE (XEXP (x, 0)) == DImode - && GET_MODE (XEXP (XEXP (x, 0), 0)) == SImode - && CONST_INT_P (XEXP (x, 1))) - { - type = (GET_CODE (XEXP (x, 0)) == SIGN_EXTEND) - ? ADDRESS_REG_SXTW : ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = exact_log2 (INTVAL (XEXP (x, 1))); - } - /* (ashift:DI (sign_extend:DI (reg:SI)) (const_int shift)) */ - else if (GET_CODE (x) == ASHIFT - && (GET_CODE (XEXP (x, 0)) == SIGN_EXTEND - || GET_CODE (XEXP (x, 0)) == ZERO_EXTEND) - && GET_MODE (XEXP (x, 0)) == DImode - && GET_MODE (XEXP (XEXP (x, 0), 0)) == SImode - && CONST_INT_P (XEXP (x, 1))) - { - type = (GET_CODE (XEXP (x, 0)) == SIGN_EXTEND) - ? ADDRESS_REG_SXTW : ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = INTVAL (XEXP (x, 1)); - } - /* (sign_extract:DI (mult:DI (reg:DI) (const_int scale)) 32+shift 0) */ - else if ((GET_CODE (x) == SIGN_EXTRACT - || GET_CODE (x) == ZERO_EXTRACT) - && GET_MODE (x) == DImode - && GET_CODE (XEXP (x, 0)) == MULT - && GET_MODE (XEXP (XEXP (x, 0), 0)) == DImode - && CONST_INT_P (XEXP (XEXP (x, 0), 1))) - { - type = (GET_CODE (x) == SIGN_EXTRACT) - ? ADDRESS_REG_SXTW : ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = exact_log2 (INTVAL (XEXP (XEXP (x, 0), 1))); - if (INTVAL (XEXP (x, 1)) != 32 + shift - || INTVAL (XEXP (x, 2)) != 0) - shift = -1; - } - /* (and:DI (mult:DI (reg:DI) (const_int scale)) - (const_int 0xffffffff<<shift)) */ - else if (GET_CODE (x) == AND - && GET_MODE (x) == DImode - && GET_CODE (XEXP (x, 0)) == MULT - && GET_MODE (XEXP (XEXP (x, 0), 0)) == DImode - && CONST_INT_P (XEXP (XEXP (x, 0), 1)) - && CONST_INT_P (XEXP (x, 1))) - { - type = ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = exact_log2 (INTVAL (XEXP (XEXP (x, 0), 1))); - if (INTVAL (XEXP (x, 1)) != (HOST_WIDE_INT)0xffffffff << shift) - shift = -1; - } - /* (sign_extract:DI (ashift:DI (reg:DI) (const_int shift)) 32+shift 0) */ - else if ((GET_CODE (x) == SIGN_EXTRACT - || GET_CODE (x) == ZERO_EXTRACT) - && GET_MODE (x) == DImode - && GET_CODE (XEXP (x, 0)) == ASHIFT - && GET_MODE (XEXP (XEXP (x, 0), 0)) == DImode - && CONST_INT_P (XEXP (XEXP (x, 0), 1))) - { - type = (GET_CODE (x) == SIGN_EXTRACT) - ? ADDRESS_REG_SXTW : ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = INTVAL (XEXP (XEXP (x, 0), 1)); - if (INTVAL (XEXP (x, 1)) != 32 + shift - || INTVAL (XEXP (x, 2)) != 0) - shift = -1; - } - /* (and:DI (ashift:DI (reg:DI) (const_int shift)) - (const_int 0xffffffff<<shift)) */ - else if (GET_CODE (x) == AND - && GET_MODE (x) == DImode - && GET_CODE (XEXP (x, 0)) == ASHIFT - && GET_MODE (XEXP (XEXP (x, 0), 0)) == DImode - && CONST_INT_P (XEXP (XEXP (x, 0), 1)) - && CONST_INT_P (XEXP (x, 1))) - { - type = ADDRESS_REG_UXTW; - index = XEXP (XEXP (x, 0), 0); - shift = INTVAL (XEXP (XEXP (x, 0), 1)); - if (INTVAL (XEXP (x, 1)) != (HOST_WIDE_INT)0xffffffff << shift) - shift = -1; - } - /* (mult:P (reg:P) (const_int scale)) */ - else if (GET_CODE (x) == MULT - && GET_MODE (x) == Pmode - && GET_MODE (XEXP (x, 0)) == Pmode - && CONST_INT_P (XEXP (x, 1))) - { - type = ADDRESS_REG_REG; - index = XEXP (x, 0); - shift = exact_log2 (INTVAL (XEXP (x, 1))); - } - /* (ashift:P (reg:P) (const_int shift)) */ - else if (GET_CODE (x) == ASHIFT - && GET_MODE (x) == Pmode - && GET_MODE (XEXP (x, 0)) == Pmode - && CONST_INT_P (XEXP (x, 1))) - { - type = ADDRESS_REG_REG; - index = XEXP (x, 0); - shift = INTVAL (XEXP (x, 1)); - } - else - return false; - - if (GET_CODE (index) == SUBREG) - index = SUBREG_REG (index); - - if ((shift == 0 || - (shift > 0 && shift <= 3 - && (1 << shift) == GET_MODE_SIZE (mode))) - && REG_P (index) - && aarch64_regno_ok_for_index_p (REGNO (index), strict_p)) - { - info->type = type; - info->offset = index; - info->shift = shift; - return true; - } - - return false; -} - -static inline bool -offset_7bit_signed_scaled_p (enum machine_mode mode, HOST_WIDE_INT offset) -{ - return (offset >= -64 * GET_MODE_SIZE (mode) - && offset < 64 * GET_MODE_SIZE (mode) - && offset % GET_MODE_SIZE (mode) == 0); -} - -static inline bool -offset_9bit_signed_unscaled_p (enum machine_mode mode ATTRIBUTE_UNUSED, - HOST_WIDE_INT offset) -{ - return offset >= -256 && offset < 256; -} - -static inline bool -offset_12bit_unsigned_scaled_p (enum machine_mode mode, HOST_WIDE_INT offset) -{ - return (offset >= 0 - && offset < 4096 * GET_MODE_SIZE (mode) - && offset % GET_MODE_SIZE (mode) == 0); -} - -/* Return true if X is a valid address for machine mode MODE. If it is, - fill in INFO appropriately. STRICT_P is true if REG_OK_STRICT is in - effect. OUTER_CODE is PARALLEL for a load/store pair. */ - -static bool -aarch64_classify_address (struct aarch64_address_info *info, - rtx x, enum machine_mode mode, - RTX_CODE outer_code, bool strict_p) -{ - enum rtx_code code = GET_CODE (x); - rtx op0, op1; - bool allow_reg_index_p = - outer_code != PARALLEL && GET_MODE_SIZE(mode) != 16; - - /* Don't support anything other than POST_INC or REG addressing for - AdvSIMD. */ - if (aarch64_vector_mode_p (mode) - && (code != POST_INC && code != REG)) - return false; - - switch (code) - { - case REG: - case SUBREG: - info->type = ADDRESS_REG_IMM; - info->base = x; - info->offset = const0_rtx; - return aarch64_base_register_rtx_p (x, strict_p); - - case PLUS: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - if (GET_MODE_SIZE (mode) != 0 - && CONST_INT_P (op1) - && aarch64_base_register_rtx_p (op0, strict_p)) - { - HOST_WIDE_INT offset = INTVAL (op1); - - info->type = ADDRESS_REG_IMM; - info->base = op0; - info->offset = op1; - - /* TImode and TFmode values are allowed in both pairs of X - registers and individual Q registers. The available - address modes are: - X,X: 7-bit signed scaled offset - Q: 9-bit signed offset - We conservatively require an offset representable in either mode. - */ - if (mode == TImode || mode == TFmode) - return (offset_7bit_signed_scaled_p (mode, offset) - && offset_9bit_signed_unscaled_p (mode, offset)); - - if (outer_code == PARALLEL) - return ((GET_MODE_SIZE (mode) == 4 || GET_MODE_SIZE (mode) == 8) - && offset_7bit_signed_scaled_p (mode, offset)); - else - return (offset_9bit_signed_unscaled_p (mode, offset) - || offset_12bit_unsigned_scaled_p (mode, offset)); - } - - if (allow_reg_index_p) - { - /* Look for base + (scaled/extended) index register. */ - if (aarch64_base_register_rtx_p (op0, strict_p) - && aarch64_classify_index (info, op1, mode, strict_p)) - { - info->base = op0; - return true; - } - if (aarch64_base_register_rtx_p (op1, strict_p) - && aarch64_classify_index (info, op0, mode, strict_p)) - { - info->base = op1; - return true; - } - } - - return false; - - case POST_INC: - case POST_DEC: - case PRE_INC: - case PRE_DEC: - info->type = ADDRESS_REG_WB; - info->base = XEXP (x, 0); - info->offset = NULL_RTX; - return aarch64_base_register_rtx_p (info->base, strict_p); - - case POST_MODIFY: - case PRE_MODIFY: - info->type = ADDRESS_REG_WB; - info->base = XEXP (x, 0); - if (GET_CODE (XEXP (x, 1)) == PLUS - && CONST_INT_P (XEXP (XEXP (x, 1), 1)) - && rtx_equal_p (XEXP (XEXP (x, 1), 0), info->base) - && aarch64_base_register_rtx_p (info->base, strict_p)) - { - HOST_WIDE_INT offset; - info->offset = XEXP (XEXP (x, 1), 1); - offset = INTVAL (info->offset); - - /* TImode and TFmode values are allowed in both pairs of X - registers and individual Q registers. The available - address modes are: - X,X: 7-bit signed scaled offset - Q: 9-bit signed offset - We conservatively require an offset representable in either mode. - */ - if (mode == TImode || mode == TFmode) - return (offset_7bit_signed_scaled_p (mode, offset) - && offset_9bit_signed_unscaled_p (mode, offset)); - - if (outer_code == PARALLEL) - return ((GET_MODE_SIZE (mode) == 4 || GET_MODE_SIZE (mode) == 8) - && offset_7bit_signed_scaled_p (mode, offset)); - else - return offset_9bit_signed_unscaled_p (mode, offset); - } - return false; - - case CONST: - case SYMBOL_REF: - case LABEL_REF: - /* load literal: pc-relative constant pool entry. Only supported - for SI mode or larger. */ - info->type = ADDRESS_SYMBOLIC; - if (outer_code != PARALLEL && GET_MODE_SIZE (mode) >= 4) - { - rtx sym, addend; - - split_const (x, &sym, &addend); - return (GET_CODE (sym) == LABEL_REF - || (GET_CODE (sym) == SYMBOL_REF - && CONSTANT_POOL_ADDRESS_P (sym))); - } - return false; - - case LO_SUM: - info->type = ADDRESS_LO_SUM; - info->base = XEXP (x, 0); - info->offset = XEXP (x, 1); - if (allow_reg_index_p - && aarch64_base_register_rtx_p (info->base, strict_p)) - { - rtx sym, offs; - split_const (info->offset, &sym, &offs); - if (GET_CODE (sym) == SYMBOL_REF - && (aarch64_classify_symbol (sym, SYMBOL_CONTEXT_MEM) - == SYMBOL_SMALL_ABSOLUTE)) - { - /* The symbol and offset must be aligned to the access size. */ - unsigned int align; - unsigned int ref_size; - - if (CONSTANT_POOL_ADDRESS_P (sym)) - align = GET_MODE_ALIGNMENT (get_pool_mode (sym)); - else if (TREE_CONSTANT_POOL_ADDRESS_P (sym)) - { - tree exp = SYMBOL_REF_DECL (sym); - align = TYPE_ALIGN (TREE_TYPE (exp)); - align = CONSTANT_ALIGNMENT (exp, align); - } - else if (SYMBOL_REF_DECL (sym)) - align = DECL_ALIGN (SYMBOL_REF_DECL (sym)); - else - align = BITS_PER_UNIT; - - ref_size = GET_MODE_SIZE (mode); - if (ref_size == 0) - ref_size = GET_MODE_SIZE (DImode); - - return ((INTVAL (offs) & (ref_size - 1)) == 0 - && ((align / BITS_PER_UNIT) & (ref_size - 1)) == 0); - } - } - return false; - - default: - return false; - } -} - -bool -aarch64_symbolic_address_p (rtx x) -{ - rtx offset; - - split_const (x, &x, &offset); - return GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF; -} - -/* Classify the base of symbolic expression X, given that X appears in - context CONTEXT. */ -static enum aarch64_symbol_type -aarch64_classify_symbolic_expression (rtx x, enum aarch64_symbol_context context) -{ - rtx offset; - split_const (x, &x, &offset); - return aarch64_classify_symbol (x, context); -} - - -/* Return TRUE if X is a legitimate address for accessing memory in - mode MODE. */ -static bool -aarch64_legitimate_address_hook_p (enum machine_mode mode, rtx x, bool strict_p) -{ - struct aarch64_address_info addr; - - return aarch64_classify_address (&addr, x, mode, MEM, strict_p); -} - -/* Return TRUE if X is a legitimate address for accessing memory in - mode MODE. OUTER_CODE will be PARALLEL if this is a load/store - pair operation. */ -bool -aarch64_legitimate_address_p (enum machine_mode mode, rtx x, - RTX_CODE outer_code, bool strict_p) -{ - struct aarch64_address_info addr; - - return aarch64_classify_address (&addr, x, mode, outer_code, strict_p); -} - -/* Return TRUE if rtx X is immediate constant 0.0 */ -bool -aarch64_float_const_zero_rtx_p (rtx x) -{ - REAL_VALUE_TYPE r; - - if (GET_MODE (x) == VOIDmode) - return false; - - REAL_VALUE_FROM_CONST_DOUBLE (r, x); - if (REAL_VALUE_MINUS_ZERO (r)) - return !HONOR_SIGNED_ZEROS (GET_MODE (x)); - return REAL_VALUES_EQUAL (r, dconst0); -} - -/* Return the fixed registers used for condition codes. */ - -static bool -aarch64_fixed_condition_code_regs (unsigned int *p1, unsigned int *p2) -{ - *p1 = CC_REGNUM; - *p2 = INVALID_REGNUM; - return true; -} - -enum machine_mode -aarch64_select_cc_mode (RTX_CODE code, rtx x, rtx y) -{ - /* All floating point compares return CCFP if it is an equality - comparison, and CCFPE otherwise. */ - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_FLOAT) - { - switch (code) - { - case EQ: - case NE: - case UNORDERED: - case ORDERED: - case UNLT: - case UNLE: - case UNGT: - case UNGE: - case UNEQ: - case LTGT: - return CCFPmode; - - case LT: - case LE: - case GT: - case GE: - return CCFPEmode; - - default: - gcc_unreachable (); - } - } - - if ((GET_MODE (x) == SImode || GET_MODE (x) == DImode) - && y == const0_rtx - && (code == EQ || code == NE || code == LT || code == GE) - && (GET_CODE (x) == PLUS || GET_CODE (x) == MINUS || GET_CODE (x) == AND)) - return CC_NZmode; - - /* A compare with a shifted operand. Because of canonicalization, - the comparison will have to be swapped when we emit the assembly - code. */ - if ((GET_MODE (x) == SImode || GET_MODE (x) == DImode) - && (GET_CODE (y) == REG || GET_CODE (y) == SUBREG) - && (GET_CODE (x) == ASHIFT || GET_CODE (x) == ASHIFTRT - || GET_CODE (x) == LSHIFTRT - || GET_CODE (x) == ZERO_EXTEND || GET_CODE (x) == SIGN_EXTEND)) - return CC_SWPmode; - - /* A compare of a mode narrower than SI mode against zero can be done - by extending the value in the comparison. */ - if ((GET_MODE (x) == QImode || GET_MODE (x) == HImode) - && y == const0_rtx) - /* Only use sign-extension if we really need it. */ - return ((code == GT || code == GE || code == LE || code == LT) - ? CC_SESWPmode : CC_ZESWPmode); - - /* For everything else, return CCmode. */ - return CCmode; -} - -static unsigned -aarch64_get_condition_code (rtx x) -{ - enum machine_mode mode = GET_MODE (XEXP (x, 0)); - enum rtx_code comp_code = GET_CODE (x); - - if (GET_MODE_CLASS (mode) != MODE_CC) - mode = SELECT_CC_MODE (comp_code, XEXP (x, 0), XEXP (x, 1)); - - switch (mode) - { - case CCFPmode: - case CCFPEmode: - switch (comp_code) - { - case GE: return AARCH64_GE; - case GT: return AARCH64_GT; - case LE: return AARCH64_LS; - case LT: return AARCH64_MI; - case NE: return AARCH64_NE; - case EQ: return AARCH64_EQ; - case ORDERED: return AARCH64_VC; - case UNORDERED: return AARCH64_VS; - case UNLT: return AARCH64_LT; - case UNLE: return AARCH64_LE; - case UNGT: return AARCH64_HI; - case UNGE: return AARCH64_PL; - default: gcc_unreachable (); - } - break; - - case CCmode: - switch (comp_code) - { - case NE: return AARCH64_NE; - case EQ: return AARCH64_EQ; - case GE: return AARCH64_GE; - case GT: return AARCH64_GT; - case LE: return AARCH64_LE; - case LT: return AARCH64_LT; - case GEU: return AARCH64_CS; - case GTU: return AARCH64_HI; - case LEU: return AARCH64_LS; - case LTU: return AARCH64_CC; - default: gcc_unreachable (); - } - break; - - case CC_SWPmode: - case CC_ZESWPmode: - case CC_SESWPmode: - switch (comp_code) - { - case NE: return AARCH64_NE; - case EQ: return AARCH64_EQ; - case GE: return AARCH64_LE; - case GT: return AARCH64_LT; - case LE: return AARCH64_GE; - case LT: return AARCH64_GT; - case GEU: return AARCH64_LS; - case GTU: return AARCH64_CC; - case LEU: return AARCH64_CS; - case LTU: return AARCH64_HI; - default: gcc_unreachable (); - } - break; - - case CC_NZmode: - switch (comp_code) - { - case NE: return AARCH64_NE; - case EQ: return AARCH64_EQ; - case GE: return AARCH64_PL; - case LT: return AARCH64_MI; - default: gcc_unreachable (); - } - break; - - default: - gcc_unreachable (); - break; - } -} - -static unsigned -bit_count (unsigned HOST_WIDE_INT value) -{ - unsigned count = 0; - - while (value) - { - count++; - value &= value - 1; - } - - return count; -} - -void -aarch64_print_operand (FILE *f, rtx x, char code) -{ - switch (code) - { - case 'e': - /* Print the sign/zero-extend size as a character 8->b, 16->h, 32->w. */ - { - int n; - - if (GET_CODE (x) != CONST_INT - || (n = exact_log2 (INTVAL (x) & ~7)) <= 0) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - switch (n) - { - case 3: - fputc ('b', f); - break; - case 4: - fputc ('h', f); - break; - case 5: - fputc ('w', f); - break; - default: - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - } - break; - - case 'p': - { - int n; - - /* Print N such that 2^N == X. */ - if (GET_CODE (x) != CONST_INT || (n = exact_log2 (INTVAL (x))) < 0) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - asm_fprintf (f, "%d", n); - } - break; - - case 'P': - /* Print the number of non-zero bits in X (a const_int). */ - if (GET_CODE (x) != CONST_INT) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - asm_fprintf (f, "%u", bit_count (INTVAL (x))); - break; - - case 'H': - /* Print the higher numbered register of a pair (TImode) of regs. */ - if (GET_CODE (x) != REG || !GP_REGNUM_P (REGNO (x) + 1)) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - asm_fprintf (f, "%s", reg_names [REGNO (x) + 1]); - break; - - case 'Q': - /* Print the least significant register of a pair (TImode) of regs. */ - if (GET_CODE (x) != REG || !GP_REGNUM_P (REGNO (x) + 1)) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - asm_fprintf (f, "%s", reg_names [REGNO (x) + (WORDS_BIG_ENDIAN ? 1 : 0)]); - break; - - case 'R': - /* Print the most significant register of a pair (TImode) of regs. */ - if (GET_CODE (x) != REG || !GP_REGNUM_P (REGNO (x) + 1)) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - asm_fprintf (f, "%s", reg_names [REGNO (x) + (WORDS_BIG_ENDIAN ? 0 : 1)]); - break; - - case 'm': - /* Print a condition (eq, ne, etc). */ - - /* CONST_TRUE_RTX means always -- that's the default. */ - if (x == const_true_rtx) - return; - - if (!COMPARISON_P (x)) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - fputs (aarch64_condition_codes[aarch64_get_condition_code (x)], f); - break; - - case 'M': - /* Print the inverse of a condition (eq <-> ne, etc). */ - - /* CONST_TRUE_RTX means never -- that's the default. */ - if (x == const_true_rtx) - { - fputs ("nv", f); - return; - } - - if (!COMPARISON_P (x)) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - - fputs (aarch64_condition_codes[AARCH64_INVERSE_CONDITION_CODE - (aarch64_get_condition_code (x))], f); - break; - - case 'b': - case 'h': - case 's': - case 'd': - case 'q': - /* Print a scalar FP/SIMD register name. */ - if (!REG_P (x) || !FP_REGNUM_P (REGNO (x))) - { - output_operand_lossage ("incompatible floating point / vector register operand for '%%%c'", code); - return; - } - asm_fprintf (f, "%s%c%d", REGISTER_PREFIX, code, REGNO (x) - V0_REGNUM); - break; - - case 'S': - case 'T': - case 'U': - case 'V': - /* Print the first FP/SIMD register name in a list. */ - if (!REG_P (x) || !FP_REGNUM_P (REGNO (x))) - { - output_operand_lossage ("incompatible floating point / vector register operand for '%%%c'", code); - return; - } - asm_fprintf (f, "%sv%d", REGISTER_PREFIX, - REGNO (x) - V0_REGNUM + (code - 'S')); - break; - - case 'X': - /* Print integer constant in hex. */ - if (GET_CODE (x) != CONST_INT) - { - output_operand_lossage ("invalid operand for '%%%c'", code); - return; - } - asm_fprintf (f, "0x%x", UINTVAL (x)); - break; - - case 'w': - case 'x': - /* Print a general register name or the zero register (32-bit or - 64-bit). */ - if (x == const0_rtx - || (CONST_DOUBLE_P (x) && aarch64_float_const_zero_rtx_p (x))) - { - asm_fprintf (f, "%s%czr", REGISTER_PREFIX, code); - break; - } - - if (REG_P (x) && GP_REGNUM_P (REGNO (x))) - { - asm_fprintf (f, "%s%c%d", REGISTER_PREFIX, code, - REGNO (x) - R0_REGNUM); - break; - } - - if (REG_P (x) && REGNO (x) == SP_REGNUM) - { - asm_fprintf (f, "%s%ssp", REGISTER_PREFIX, code == 'w' ? "w" : ""); - break; - } - - /* Fall through */ - - case 0: - /* Print a normal operand, if it's a general register, then we - assume DImode. */ - if (x == NULL) - { - output_operand_lossage ("missing operand"); - return; - } - - switch (GET_CODE (x)) - { - case REG: - asm_fprintf (f, "%s", reg_names [REGNO (x)]); - break; - - case MEM: - aarch64_memory_reference_mode = GET_MODE (x); - output_address (XEXP (x, 0)); - break; - - case LABEL_REF: - case SYMBOL_REF: - output_addr_const (asm_out_file, x); - break; - - case CONST_INT: - asm_fprintf (f, "%wd", INTVAL (x)); - break; - - case CONST_VECTOR: - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_VECTOR_INT) - { - gcc_assert (aarch64_const_vec_all_same_int_p (x, - HOST_WIDE_INT_MIN, - HOST_WIDE_INT_MAX)); - asm_fprintf (f, "%wd", INTVAL (CONST_VECTOR_ELT (x, 0))); - } - else if (aarch64_simd_imm_zero_p (x, GET_MODE (x))) - { - fputc ('0', f); - } - else - gcc_unreachable (); - break; - - case CONST_DOUBLE: - /* CONST_DOUBLE can represent a double-width integer. - In this case, the mode of x is VOIDmode. */ - if (GET_MODE (x) == VOIDmode) - ; /* Do Nothing. */ - else if (aarch64_float_const_zero_rtx_p (x)) - { - fputc ('0', f); - break; - } - else if (aarch64_float_const_representable_p (x)) - { -#define buf_size 20 - char float_buf[buf_size] = {'\0'}; - REAL_VALUE_TYPE r; - REAL_VALUE_FROM_CONST_DOUBLE (r, x); - real_to_decimal_for_mode (float_buf, &r, - buf_size, buf_size, - 1, GET_MODE (x)); - asm_fprintf (asm_out_file, "%s", float_buf); - break; -#undef buf_size - } - output_operand_lossage ("invalid constant"); - return; - default: - output_operand_lossage ("invalid operand"); - return; - } - break; - - case 'A': - if (GET_CODE (x) == HIGH) - x = XEXP (x, 0); - - switch (aarch64_classify_symbolic_expression (x, SYMBOL_CONTEXT_ADR)) - { - case SYMBOL_SMALL_GOT: - asm_fprintf (asm_out_file, ":got:"); - break; - - case SYMBOL_SMALL_TLSGD: - asm_fprintf (asm_out_file, ":tlsgd:"); - break; - - case SYMBOL_SMALL_TLSDESC: - asm_fprintf (asm_out_file, ":tlsdesc:"); - break; - - case SYMBOL_SMALL_GOTTPREL: - asm_fprintf (asm_out_file, ":gottprel:"); - break; - - case SYMBOL_SMALL_TPREL: - asm_fprintf (asm_out_file, ":tprel:"); - break; - - default: - break; - } - output_addr_const (asm_out_file, x); - break; - - case 'L': - switch (aarch64_classify_symbolic_expression (x, SYMBOL_CONTEXT_ADR)) - { - case SYMBOL_SMALL_GOT: - asm_fprintf (asm_out_file, ":lo12:"); - break; - - case SYMBOL_SMALL_TLSGD: - asm_fprintf (asm_out_file, ":tlsgd_lo12:"); - break; - - case SYMBOL_SMALL_TLSDESC: - asm_fprintf (asm_out_file, ":tlsdesc_lo12:"); - break; - - case SYMBOL_SMALL_GOTTPREL: - asm_fprintf (asm_out_file, ":gottprel_lo12:"); - break; - - case SYMBOL_SMALL_TPREL: - asm_fprintf (asm_out_file, ":tprel_lo12_nc:"); - break; - - default: - break; - } - output_addr_const (asm_out_file, x); - break; - - case 'G': - - switch (aarch64_classify_symbolic_expression (x, SYMBOL_CONTEXT_ADR)) - { - case SYMBOL_SMALL_TPREL: - asm_fprintf (asm_out_file, ":tprel_hi12:"); - break; - default: - break; - } - output_addr_const (asm_out_file, x); - break; - - default: - output_operand_lossage ("invalid operand prefix '%%%c'", code); - return; - } -} - -void -aarch64_print_operand_address (FILE *f, rtx x) -{ - struct aarch64_address_info addr; - - if (aarch64_classify_address (&addr, x, aarch64_memory_reference_mode, - MEM, true)) - switch (addr.type) - { - case ADDRESS_REG_IMM: - if (addr.offset == const0_rtx) - asm_fprintf (f, "[%s]", reg_names [REGNO (addr.base)]); - else - asm_fprintf (f, "[%s,%wd]", reg_names [REGNO (addr.base)], - INTVAL (addr.offset)); - return; - - case ADDRESS_REG_REG: - if (addr.shift == 0) - asm_fprintf (f, "[%s,%s]", reg_names [REGNO (addr.base)], - reg_names [REGNO (addr.offset)]); - else - asm_fprintf (f, "[%s,%s,lsl %u]", reg_names [REGNO (addr.base)], - reg_names [REGNO (addr.offset)], addr.shift); - return; - - case ADDRESS_REG_UXTW: - if (addr.shift == 0) - asm_fprintf (f, "[%s,w%d,uxtw]", reg_names [REGNO (addr.base)], - REGNO (addr.offset) - R0_REGNUM); - else - asm_fprintf (f, "[%s,w%d,uxtw %u]", reg_names [REGNO (addr.base)], - REGNO (addr.offset) - R0_REGNUM, addr.shift); - return; - - case ADDRESS_REG_SXTW: - if (addr.shift == 0) - asm_fprintf (f, "[%s,w%d,sxtw]", reg_names [REGNO (addr.base)], - REGNO (addr.offset) - R0_REGNUM); - else - asm_fprintf (f, "[%s,w%d,sxtw %u]", reg_names [REGNO (addr.base)], - REGNO (addr.offset) - R0_REGNUM, addr.shift); - return; - - case ADDRESS_REG_WB: - switch (GET_CODE (x)) - { - case PRE_INC: - asm_fprintf (f, "[%s,%d]!", reg_names [REGNO (addr.base)], - GET_MODE_SIZE (aarch64_memory_reference_mode)); - return; - case POST_INC: - asm_fprintf (f, "[%s],%d", reg_names [REGNO (addr.base)], - GET_MODE_SIZE (aarch64_memory_reference_mode)); - return; - case PRE_DEC: - asm_fprintf (f, "[%s,-%d]!", reg_names [REGNO (addr.base)], - GET_MODE_SIZE (aarch64_memory_reference_mode)); - return; - case POST_DEC: - asm_fprintf (f, "[%s],-%d", reg_names [REGNO (addr.base)], - GET_MODE_SIZE (aarch64_memory_reference_mode)); - return; - case PRE_MODIFY: - asm_fprintf (f, "[%s,%wd]!", reg_names [REGNO (addr.base)], - INTVAL (addr.offset)); - return; - case POST_MODIFY: - asm_fprintf (f, "[%s],%wd", reg_names [REGNO (addr.base)], - INTVAL (addr.offset)); - return; - default: - break; - } - break; - - case ADDRESS_LO_SUM: - asm_fprintf (f, "[%s,#:lo12:", reg_names [REGNO (addr.base)]); - output_addr_const (f, addr.offset); - asm_fprintf (f, "]"); - return; - - case ADDRESS_SYMBOLIC: - break; - } - - output_addr_const (f, x); -} - -void -aarch64_function_profiler (FILE *f ATTRIBUTE_UNUSED, - int labelno ATTRIBUTE_UNUSED) -{ - sorry ("function profiling"); -} - -bool -aarch64_label_mentioned_p (rtx x) -{ - const char *fmt; - int i; - - if (GET_CODE (x) == LABEL_REF) - return true; - - /* UNSPEC_TLS entries for a symbol include a LABEL_REF for the - referencing instruction, but they are constant offsets, not - symbols. */ - if (GET_CODE (x) == UNSPEC && XINT (x, 1) == UNSPEC_TLS) - return false; - - fmt = GET_RTX_FORMAT (GET_CODE (x)); - for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--) - { - if (fmt[i] == 'E') - { - int j; - - for (j = XVECLEN (x, i) - 1; j >= 0; j--) - if (aarch64_label_mentioned_p (XVECEXP (x, i, j))) - return 1; - } - else if (fmt[i] == 'e' && aarch64_label_mentioned_p (XEXP (x, i))) - return 1; - } - - return 0; -} - -/* Implement REGNO_REG_CLASS. */ - -enum reg_class -aarch64_regno_regclass (unsigned regno) -{ - if (GP_REGNUM_P (regno)) - return CORE_REGS; - - if (regno == SP_REGNUM) - return STACK_REG; - - if (regno == FRAME_POINTER_REGNUM - || regno == ARG_POINTER_REGNUM) - return CORE_REGS; - - if (FP_REGNUM_P (regno)) - return FP_LO_REGNUM_P (regno) ? FP_LO_REGS : FP_REGS; - - return NO_REGS; -} - -/* Try a machine-dependent way of reloading an illegitimate address - operand. If we find one, push the reload and return the new rtx. */ - -rtx -aarch64_legitimize_reload_address (rtx *x_p, - enum machine_mode mode, - int opnum, int type, - int ind_levels ATTRIBUTE_UNUSED) -{ - rtx x = *x_p; - - /* Do not allow mem (plus (reg, const)) if vector mode. */ - if (aarch64_vector_mode_p (mode) - && GET_CODE (x) == PLUS - && REG_P (XEXP (x, 0)) - && CONST_INT_P (XEXP (x, 1))) - { - rtx orig_rtx = x; - x = copy_rtx (x); - push_reload (orig_rtx, NULL_RTX, x_p, NULL, - BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0, - opnum, (enum reload_type) type); - return x; - } - - /* We must recognize output that we have already generated ourselves. */ - if (GET_CODE (x) == PLUS - && GET_CODE (XEXP (x, 0)) == PLUS - && REG_P (XEXP (XEXP (x, 0), 0)) - && CONST_INT_P (XEXP (XEXP (x, 0), 1)) - && CONST_INT_P (XEXP (x, 1))) - { - push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL, - BASE_REG_CLASS, GET_MODE (x), VOIDmode, 0, 0, - opnum, (enum reload_type) type); - return x; - } - - /* We wish to handle large displacements off a base register by splitting - the addend across an add and the mem insn. This can cut the number of - extra insns needed from 3 to 1. It is only useful for load/store of a - single register with 12 bit offset field. */ - if (GET_CODE (x) == PLUS - && REG_P (XEXP (x, 0)) - && CONST_INT_P (XEXP (x, 1)) - && HARD_REGISTER_P (XEXP (x, 0)) - && mode != TImode - && mode != TFmode - && aarch64_regno_ok_for_base_p (REGNO (XEXP (x, 0)), true)) - { - HOST_WIDE_INT val = INTVAL (XEXP (x, 1)); - HOST_WIDE_INT low = val & 0xfff; - HOST_WIDE_INT high = val - low; - HOST_WIDE_INT offs; - rtx cst; - - /* Reload non-zero BLKmode offsets. This is because we cannot ascertain - BLKmode alignment. */ - if (GET_MODE_SIZE (mode) == 0) - return NULL_RTX; - - offs = low % GET_MODE_SIZE (mode); - - /* Align misaligned offset by adjusting high part to compensate. */ - if (offs != 0) - { - if (aarch64_uimm12_shift (high + offs)) - { - /* Align down. */ - low = low - offs; - high = high + offs; - } - else - { - /* Align up. */ - offs = GET_MODE_SIZE (mode) - offs; - low = low + offs; - high = high + (low & 0x1000) - offs; - low &= 0xfff; - } - } - - /* Check for overflow. */ - if (high + low != val) - return NULL_RTX; - - cst = GEN_INT (high); - if (!aarch64_uimm12_shift (high)) - cst = force_const_mem (Pmode, cst); - - /* Reload high part into base reg, leaving the low part - in the mem instruction. */ - x = gen_rtx_PLUS (Pmode, - gen_rtx_PLUS (Pmode, XEXP (x, 0), cst), - GEN_INT (low)); - - push_reload (XEXP (x, 0), NULL_RTX, &XEXP (x, 0), NULL, - BASE_REG_CLASS, Pmode, VOIDmode, 0, 0, - opnum, (enum reload_type) type); - return x; - } - - return NULL_RTX; -} - - -static reg_class_t -aarch64_secondary_reload (bool in_p ATTRIBUTE_UNUSED, rtx x, - reg_class_t rclass, - enum machine_mode mode, - secondary_reload_info *sri) -{ - /* Address expressions of the form PLUS (SP, large_offset) need two - scratch registers, one for the constant, and one for holding a - copy of SP, since SP cannot be used on the RHS of an add-reg - instruction. */ - if (mode == DImode - && GET_CODE (x) == PLUS - && XEXP (x, 0) == stack_pointer_rtx - && CONST_INT_P (XEXP (x, 1)) - && !aarch64_uimm12_shift (INTVAL (XEXP (x, 1)))) - { - sri->icode = CODE_FOR_reload_sp_immediate; - return NO_REGS; - } - - /* Without the TARGET_SIMD instructions we cannot move a Q register - to a Q register directly. We need a scratch. */ - if (REG_P (x) && (mode == TFmode || mode == TImode) && mode == GET_MODE (x) - && FP_REGNUM_P (REGNO (x)) && !TARGET_SIMD - && reg_class_subset_p (rclass, FP_REGS)) - { - if (mode == TFmode) - sri->icode = CODE_FOR_aarch64_reload_movtf; - else if (mode == TImode) - sri->icode = CODE_FOR_aarch64_reload_movti; - return NO_REGS; - } - - /* A TFmode or TImode memory access should be handled via an FP_REGS - because AArch64 has richer addressing modes for LDR/STR instructions - than LDP/STP instructions. */ - if (!TARGET_GENERAL_REGS_ONLY && rclass == CORE_REGS - && GET_MODE_SIZE (mode) == 16 && MEM_P (x)) - return FP_REGS; - - if (rclass == FP_REGS && (mode == TImode || mode == TFmode) && CONSTANT_P(x)) - return CORE_REGS; - - return NO_REGS; -} - -static bool -aarch64_can_eliminate (const int from, const int to) -{ - /* If we need a frame pointer, we must eliminate FRAME_POINTER_REGNUM into - HARD_FRAME_POINTER_REGNUM and not into STACK_POINTER_REGNUM. */ - - if (frame_pointer_needed) - { - if (from == ARG_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM) - return true; - if (from == ARG_POINTER_REGNUM && to == STACK_POINTER_REGNUM) - return false; - if (from == FRAME_POINTER_REGNUM && to == STACK_POINTER_REGNUM - && !cfun->calls_alloca) - return true; - if (from == FRAME_POINTER_REGNUM && to == HARD_FRAME_POINTER_REGNUM) - return true; - return false; - } - else - { - /* If we decided that we didn't need a leaf frame pointer but then used - LR in the function, then we'll want a frame pointer after all, so - prevent this elimination to ensure a frame pointer is used. - - NOTE: the original value of flag_omit_frame_pointer gets trashed - IFF flag_omit_leaf_frame_pointer is true, so we check the value - of faked_omit_frame_pointer here (which is true when we always - wish to keep non-leaf frame pointers but only wish to keep leaf frame - pointers when LR is clobbered). */ - if (from == FRAME_POINTER_REGNUM && to == STACK_POINTER_REGNUM - && df_regs_ever_live_p (LR_REGNUM) - && faked_omit_frame_pointer) - return false; - } - - return true; -} - -HOST_WIDE_INT -aarch64_initial_elimination_offset (unsigned from, unsigned to) -{ - HOST_WIDE_INT frame_size; - HOST_WIDE_INT offset; - - aarch64_layout_frame (); - frame_size = (get_frame_size () + cfun->machine->frame.saved_regs_size - + crtl->outgoing_args_size - + cfun->machine->saved_varargs_size); - - frame_size = AARCH64_ROUND_UP (frame_size, STACK_BOUNDARY / BITS_PER_UNIT); - offset = frame_size; - - if (to == HARD_FRAME_POINTER_REGNUM) - { - if (from == ARG_POINTER_REGNUM) - return offset - crtl->outgoing_args_size; - - if (from == FRAME_POINTER_REGNUM) - return cfun->machine->frame.saved_regs_size; - } - - if (to == STACK_POINTER_REGNUM) - { - if (from == FRAME_POINTER_REGNUM) - { - HOST_WIDE_INT elim = crtl->outgoing_args_size - + cfun->machine->frame.saved_regs_size - - cfun->machine->frame.fp_lr_offset; - elim = AARCH64_ROUND_UP (elim, STACK_BOUNDARY / BITS_PER_UNIT); - return elim; - } - } - - return offset; -} - - -/* Implement RETURN_ADDR_RTX. We do not support moving back to a - previous frame. */ - -rtx -aarch64_return_addr (int count, rtx frame ATTRIBUTE_UNUSED) -{ - if (count != 0) - return const0_rtx; - return get_hard_reg_initial_val (Pmode, LR_REGNUM); -} - - -static void -aarch64_asm_trampoline_template (FILE *f) -{ - asm_fprintf (f, "\tldr\t%s, .+16\n", reg_names [IP1_REGNUM]); - asm_fprintf (f, "\tldr\t%s, .+20\n", reg_names [STATIC_CHAIN_REGNUM]); - asm_fprintf (f, "\tbr\t%s\n", reg_names [IP1_REGNUM]); - assemble_aligned_integer (4, const0_rtx); - assemble_aligned_integer (UNITS_PER_WORD, const0_rtx); - assemble_aligned_integer (UNITS_PER_WORD, const0_rtx); -} - -unsigned -aarch64_trampoline_size (void) -{ - return 32; /* 3 insns + padding + 2 dwords. */ -} - -static void -aarch64_trampoline_init (rtx m_tramp, tree fndecl, rtx chain_value) -{ - rtx fnaddr, mem, a_tramp; - - /* Don't need to copy the trailing D-words, we fill those in below. */ - emit_block_move (m_tramp, assemble_trampoline_template (), - GEN_INT (TRAMPOLINE_SIZE - 16), BLOCK_OP_NORMAL); - mem = adjust_address (m_tramp, DImode, 16); - fnaddr = XEXP (DECL_RTL (fndecl), 0); - emit_move_insn (mem, fnaddr); - - mem = adjust_address (m_tramp, DImode, 24); - emit_move_insn (mem, chain_value); - - /* XXX We should really define a "clear_cache" pattern and use - gen_clear_cache(). */ - a_tramp = XEXP (m_tramp, 0); - emit_library_call (gen_rtx_SYMBOL_REF (Pmode, "__clear_cache"), - LCT_NORMAL, VOIDmode, 2, a_tramp, Pmode, - plus_constant (Pmode, a_tramp, TRAMPOLINE_SIZE), Pmode); -} - -static unsigned char -aarch64_class_max_nregs (reg_class_t regclass, enum machine_mode mode) -{ - switch (regclass) - { - case CORE_REGS: - case POINTER_REGS: - case GENERAL_REGS: - case ALL_REGS: - case FP_REGS: - case FP_LO_REGS: - return - aarch64_vector_mode_p (mode) ? (GET_MODE_SIZE (mode) + 15) / 16 : - (GET_MODE_SIZE (mode) + 7) / 8; - case STACK_REG: - return 1; - - case NO_REGS: - return 0; - - default: - break; - } - gcc_unreachable (); -} - -static reg_class_t -aarch64_preferred_reload_class (rtx x ATTRIBUTE_UNUSED, reg_class_t regclass) -{ - return ((regclass == POINTER_REGS || regclass == STACK_REG) - ? GENERAL_REGS : regclass); -} - -void -aarch64_asm_output_labelref (FILE* f, const char *name) -{ - asm_fprintf (f, "%U%s", name); -} - -static void -aarch64_elf_asm_constructor (rtx symbol, int priority) -{ - if (priority == DEFAULT_INIT_PRIORITY) - default_ctor_section_asm_out_constructor (symbol, priority); - else - { - section *s; - char buf[18]; - snprintf (buf, sizeof (buf), ".init_array.%.5u", priority); - s = get_section (buf, SECTION_WRITE, NULL); - switch_to_section (s); - assemble_align (POINTER_SIZE); - fputs ("\t.dword\t", asm_out_file); - output_addr_const (asm_out_file, symbol); - fputc ('\n', asm_out_file); - } -} - -static void -aarch64_elf_asm_destructor (rtx symbol, int priority) -{ - if (priority == DEFAULT_INIT_PRIORITY) - default_dtor_section_asm_out_destructor (symbol, priority); - else - { - section *s; - char buf[18]; - snprintf (buf, sizeof (buf), ".fini_array.%.5u", priority); - s = get_section (buf, SECTION_WRITE, NULL); - switch_to_section (s); - assemble_align (POINTER_SIZE); - fputs ("\t.dword\t", asm_out_file); - output_addr_const (asm_out_file, symbol); - fputc ('\n', asm_out_file); - } -} - -const char* -aarch64_output_casesi (rtx *operands) -{ - char buf[100]; - char label[100]; - rtx diff_vec = PATTERN (next_real_insn (operands[2])); - int index; - static const char *const patterns[4][2] = - { - { - "ldrb\t%w3, [%0,%w1,uxtw]", - "add\t%3, %4, %w3, sxtb #2" - }, - { - "ldrh\t%w3, [%0,%w1,uxtw #1]", - "add\t%3, %4, %w3, sxth #2" - }, - { - "ldr\t%w3, [%0,%w1,uxtw #2]", - "add\t%3, %4, %w3, sxtw #2" - }, - /* We assume that DImode is only generated when not optimizing and - that we don't really need 64-bit address offsets. That would - imply an object file with 8GB of code in a single function! */ - { - "ldr\t%w3, [%0,%w1,uxtw #2]", - "add\t%3, %4, %w3, sxtw #2" - } - }; - - gcc_assert (GET_CODE (diff_vec) == ADDR_DIFF_VEC); - - index = exact_log2 (GET_MODE_SIZE (GET_MODE (diff_vec))); - - gcc_assert (index >= 0 && index <= 3); - - /* Need to implement table size reduction, by chaning the code below. */ - output_asm_insn (patterns[index][0], operands); - ASM_GENERATE_INTERNAL_LABEL (label, "Lrtx", CODE_LABEL_NUMBER (operands[2])); - snprintf (buf, sizeof (buf), - "adr\t%%4, %s", targetm.strip_name_encoding (label)); - output_asm_insn (buf, operands); - output_asm_insn (patterns[index][1], operands); - output_asm_insn ("br\t%3", operands); - assemble_label (asm_out_file, label); - return ""; -} - - -/* Return size in bits of an arithmetic operand which is shifted/scaled and - masked such that it is suitable for a UXTB, UXTH, or UXTW extend - operator. */ - -int -aarch64_uxt_size (int shift, HOST_WIDE_INT mask) -{ - if (shift >= 0 && shift <= 3) - { - int size; - for (size = 8; size <= 32; size *= 2) - { - HOST_WIDE_INT bits = ((HOST_WIDE_INT)1U << size) - 1; - if (mask == bits << shift) - return size; - } - } - return 0; -} - -static bool -aarch64_use_blocks_for_constant_p (enum machine_mode mode ATTRIBUTE_UNUSED, - const_rtx x ATTRIBUTE_UNUSED) -{ - /* We can't use blocks for constants when we're using a per-function - constant pool. */ - return false; -} - -static section * -aarch64_select_rtx_section (enum machine_mode mode ATTRIBUTE_UNUSED, - rtx x ATTRIBUTE_UNUSED, - unsigned HOST_WIDE_INT align ATTRIBUTE_UNUSED) -{ - /* Force all constant pool entries into the current function section. */ - return function_section (current_function_decl); -} - - -/* Costs. */ - -/* Helper function for rtx cost calculation. Strip a shift expression - from X. Returns the inner operand if successful, or the original - expression on failure. */ -static rtx -aarch64_strip_shift (rtx x) -{ - rtx op = x; - - if ((GET_CODE (op) == ASHIFT - || GET_CODE (op) == ASHIFTRT - || GET_CODE (op) == LSHIFTRT) - && CONST_INT_P (XEXP (op, 1))) - return XEXP (op, 0); - - if (GET_CODE (op) == MULT - && CONST_INT_P (XEXP (op, 1)) - && ((unsigned) exact_log2 (INTVAL (XEXP (op, 1)))) < 64) - return XEXP (op, 0); - - return x; -} - -/* Helper function for rtx cost calculation. Strip a shift or extend - expression from X. Returns the inner operand if successful, or the - original expression on failure. We deal with a number of possible - canonicalization variations here. */ -static rtx -aarch64_strip_shift_or_extend (rtx x) -{ - rtx op = x; - - /* Zero and sign extraction of a widened value. */ - if ((GET_CODE (op) == ZERO_EXTRACT || GET_CODE (op) == SIGN_EXTRACT) - && XEXP (op, 2) == const0_rtx - && aarch64_is_extend_from_extract (GET_MODE (op), XEXP (XEXP (op, 0), 1), - XEXP (op, 1))) - return XEXP (XEXP (op, 0), 0); - - /* It can also be represented (for zero-extend) as an AND with an - immediate. */ - if (GET_CODE (op) == AND - && GET_CODE (XEXP (op, 0)) == MULT - && CONST_INT_P (XEXP (XEXP (op, 0), 1)) - && CONST_INT_P (XEXP (op, 1)) - && aarch64_uxt_size (exact_log2 (INTVAL (XEXP (XEXP (op, 0), 1))), - INTVAL (XEXP (op, 1))) != 0) - return XEXP (XEXP (op, 0), 0); - - /* Now handle extended register, as this may also have an optional - left shift by 1..4. */ - if (GET_CODE (op) == ASHIFT - && CONST_INT_P (XEXP (op, 1)) - && ((unsigned HOST_WIDE_INT) INTVAL (XEXP (op, 1))) <= 4) - op = XEXP (op, 0); - - if (GET_CODE (op) == ZERO_EXTEND - || GET_CODE (op) == SIGN_EXTEND) - op = XEXP (op, 0); - - if (op != x) - return op; - - return aarch64_strip_shift (x); -} - -/* Calculate the cost of calculating X, storing it in *COST. Result - is true if the total cost of the operation has now been calculated. */ -static bool -aarch64_rtx_costs (rtx x, int code, int outer ATTRIBUTE_UNUSED, - int param ATTRIBUTE_UNUSED, int *cost, bool speed) -{ - rtx op0, op1; - const struct cpu_rtx_cost_table *extra_cost - = aarch64_tune_params->insn_extra_cost; - - switch (code) - { - case SET: - op0 = SET_DEST (x); - op1 = SET_SRC (x); - - switch (GET_CODE (op0)) - { - case MEM: - if (speed) - *cost += extra_cost->memory_store; - - if (op1 != const0_rtx) - *cost += rtx_cost (op1, SET, 1, speed); - return true; - - case SUBREG: - if (! REG_P (SUBREG_REG (op0))) - *cost += rtx_cost (SUBREG_REG (op0), SET, 0, speed); - /* Fall through. */ - case REG: - /* Cost is just the cost of the RHS of the set. */ - *cost += rtx_cost (op1, SET, 1, true); - return true; - - case ZERO_EXTRACT: /* Bit-field insertion. */ - case SIGN_EXTRACT: - /* Strip any redundant widening of the RHS to meet the width of - the target. */ - if (GET_CODE (op1) == SUBREG) - op1 = SUBREG_REG (op1); - if ((GET_CODE (op1) == ZERO_EXTEND - || GET_CODE (op1) == SIGN_EXTEND) - && GET_CODE (XEXP (op0, 1)) == CONST_INT - && (GET_MODE_BITSIZE (GET_MODE (XEXP (op1, 0))) - >= INTVAL (XEXP (op0, 1)))) - op1 = XEXP (op1, 0); - *cost += rtx_cost (op1, SET, 1, speed); - return true; - - default: - break; - } - return false; - - case MEM: - if (speed) - *cost += extra_cost->memory_load; - - return true; - - case NEG: - op0 = CONST0_RTX (GET_MODE (x)); - op1 = XEXP (x, 0); - goto cost_minus; - - case COMPARE: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - - if (op1 == const0_rtx - && GET_CODE (op0) == AND) - { - x = op0; - goto cost_logic; - } - - /* Comparisons can work if the order is swapped. - Canonicalization puts the more complex operation first, but - we want it in op1. */ - if (! (REG_P (op0) - || (GET_CODE (op0) == SUBREG && REG_P (SUBREG_REG (op0))))) - { - op0 = XEXP (x, 1); - op1 = XEXP (x, 0); - } - goto cost_minus; - - case MINUS: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - - cost_minus: - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT - || (GET_MODE_CLASS (GET_MODE (x)) == MODE_CC - && GET_MODE_CLASS (GET_MODE (op0)) == MODE_INT)) - { - if (op0 != const0_rtx) - *cost += rtx_cost (op0, MINUS, 0, speed); - - if (CONST_INT_P (op1)) - { - if (!aarch64_uimm12_shift (INTVAL (op1))) - *cost += rtx_cost (op1, MINUS, 1, speed); - } - else - { - op1 = aarch64_strip_shift_or_extend (op1); - *cost += rtx_cost (op1, MINUS, 1, speed); - } - return true; - } - - return false; - - case PLUS: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT) - { - if (CONST_INT_P (op1) && aarch64_uimm12_shift (INTVAL (op1))) - { - *cost += rtx_cost (op0, PLUS, 0, speed); - } - else - { - rtx new_op0 = aarch64_strip_shift_or_extend (op0); - - if (new_op0 == op0 - && GET_CODE (op0) == MULT) - { - if ((GET_CODE (XEXP (op0, 0)) == ZERO_EXTEND - && GET_CODE (XEXP (op0, 1)) == ZERO_EXTEND) - || (GET_CODE (XEXP (op0, 0)) == SIGN_EXTEND - && GET_CODE (XEXP (op0, 1)) == SIGN_EXTEND)) - { - *cost += (rtx_cost (XEXP (XEXP (op0, 0), 0), MULT, 0, - speed) - + rtx_cost (XEXP (XEXP (op0, 1), 0), MULT, 1, - speed) - + rtx_cost (op1, PLUS, 1, speed)); - if (speed) - *cost += extra_cost->int_multiply_extend_add; - return true; - } - *cost += (rtx_cost (XEXP (op0, 0), MULT, 0, speed) - + rtx_cost (XEXP (op0, 1), MULT, 1, speed) - + rtx_cost (op1, PLUS, 1, speed)); - - if (speed) - *cost += extra_cost->int_multiply_add; - } - - *cost += (rtx_cost (new_op0, PLUS, 0, speed) - + rtx_cost (op1, PLUS, 1, speed)); - } - return true; - } - - return false; - - case IOR: - case XOR: - case AND: - cost_logic: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT) - { - if (CONST_INT_P (op1) - && aarch64_bitmask_imm (INTVAL (op1), GET_MODE (x))) - { - *cost += rtx_cost (op0, AND, 0, speed); - } - else - { - if (GET_CODE (op0) == NOT) - op0 = XEXP (op0, 0); - op0 = aarch64_strip_shift (op0); - *cost += (rtx_cost (op0, AND, 0, speed) - + rtx_cost (op1, AND, 1, speed)); - } - return true; - } - return false; - - case ZERO_EXTEND: - if ((GET_MODE (x) == DImode - && GET_MODE (XEXP (x, 0)) == SImode) - || GET_CODE (XEXP (x, 0)) == MEM) - { - *cost += rtx_cost (XEXP (x, 0), ZERO_EXTEND, 0, speed); - return true; - } - return false; - - case SIGN_EXTEND: - if (GET_CODE (XEXP (x, 0)) == MEM) - { - *cost += rtx_cost (XEXP (x, 0), SIGN_EXTEND, 0, speed); - return true; - } - return false; - - case ROTATE: - if (!CONST_INT_P (XEXP (x, 1))) - *cost += COSTS_N_INSNS (2); - /* Fall through. */ - case ROTATERT: - case LSHIFTRT: - case ASHIFT: - case ASHIFTRT: - - /* Shifting by a register often takes an extra cycle. */ - if (speed && !CONST_INT_P (XEXP (x, 1))) - *cost += extra_cost->register_shift; - - *cost += rtx_cost (XEXP (x, 0), ASHIFT, 0, speed); - return true; - - case HIGH: - if (!CONSTANT_P (XEXP (x, 0))) - *cost += rtx_cost (XEXP (x, 0), HIGH, 0, speed); - return true; - - case LO_SUM: - if (!CONSTANT_P (XEXP (x, 1))) - *cost += rtx_cost (XEXP (x, 1), LO_SUM, 1, speed); - *cost += rtx_cost (XEXP (x, 0), LO_SUM, 0, speed); - return true; - - case ZERO_EXTRACT: - case SIGN_EXTRACT: - *cost += rtx_cost (XEXP (x, 0), ZERO_EXTRACT, 0, speed); - return true; - - case MULT: - op0 = XEXP (x, 0); - op1 = XEXP (x, 1); - - *cost = COSTS_N_INSNS (1); - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT) - { - if (CONST_INT_P (op1) - && exact_log2 (INTVAL (op1)) > 0) - { - *cost += rtx_cost (op0, ASHIFT, 0, speed); - return true; - } - - if ((GET_CODE (op0) == ZERO_EXTEND - && GET_CODE (op1) == ZERO_EXTEND) - || (GET_CODE (op0) == SIGN_EXTEND - && GET_CODE (op1) == SIGN_EXTEND)) - { - *cost += (rtx_cost (XEXP (op0, 0), MULT, 0, speed) - + rtx_cost (XEXP (op1, 0), MULT, 1, speed)); - if (speed) - *cost += extra_cost->int_multiply_extend; - return true; - } - - if (speed) - *cost += extra_cost->int_multiply; - } - else if (speed) - { - if (GET_MODE (x) == DFmode) - *cost += extra_cost->double_multiply; - else if (GET_MODE (x) == SFmode) - *cost += extra_cost->float_multiply; - } - - return false; /* All arguments need to be in registers. */ - - case MOD: - case UMOD: - *cost = COSTS_N_INSNS (2); - if (speed) - { - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT) - *cost += (extra_cost->int_multiply_add - + extra_cost->int_divide); - else if (GET_MODE (x) == DFmode) - *cost += (extra_cost->double_multiply - + extra_cost->double_divide); - else if (GET_MODE (x) == SFmode) - *cost += (extra_cost->float_multiply - + extra_cost->float_divide); - } - return false; /* All arguments need to be in registers. */ - - case DIV: - case UDIV: - *cost = COSTS_N_INSNS (1); - if (speed) - { - if (GET_MODE_CLASS (GET_MODE (x)) == MODE_INT) - *cost += extra_cost->int_divide; - else if (GET_MODE (x) == DFmode) - *cost += extra_cost->double_divide; - else if (GET_MODE (x) == SFmode) - *cost += extra_cost->float_divide; - } - return false; /* All arguments need to be in registers. */ - - default: - break; - } - return false; -} - -static int -aarch64_address_cost (rtx x ATTRIBUTE_UNUSED, - enum machine_mode mode ATTRIBUTE_UNUSED, - addr_space_t as ATTRIBUTE_UNUSED, bool speed ATTRIBUTE_UNUSED) -{ - enum rtx_code c = GET_CODE (x); - const struct cpu_addrcost_table *addr_cost = aarch64_tune_params->addr_cost; - - if (c == PRE_INC || c == PRE_DEC || c == PRE_MODIFY) - return addr_cost->pre_modify; - - if (c == POST_INC || c == POST_DEC || c == POST_MODIFY) - return addr_cost->post_modify; - - if (c == PLUS) - { - if (GET_CODE (XEXP (x, 1)) == CONST_INT) - return addr_cost->imm_offset; - else if (GET_CODE (XEXP (x, 0)) == MULT - || GET_CODE (XEXP (x, 0)) == ZERO_EXTEND - || GET_CODE (XEXP (x, 0)) == SIGN_EXTEND) - return addr_cost->register_extend; - - return addr_cost->register_offset; - } - else if (c == MEM || c == LABEL_REF || c == SYMBOL_REF) - return addr_cost->imm_offset; - - return 0; -} - -static int -aarch64_register_move_cost (enum machine_mode mode ATTRIBUTE_UNUSED, - reg_class_t from, reg_class_t to) -{ - const struct cpu_regmove_cost *regmove_cost - = aarch64_tune_params->regmove_cost; - - if (from == GENERAL_REGS && to == GENERAL_REGS) - return regmove_cost->GP2GP; - else if (from == GENERAL_REGS) - return regmove_cost->GP2FP; - else if (to == GENERAL_REGS) - return regmove_cost->FP2GP; - - /* When AdvSIMD instructions are disabled it is not possible to move - a 128-bit value directly between Q registers. This is handled in - secondary reload. A general register is used as a scratch to move - the upper DI value and the lower DI value is moved directly, - hence the cost is the sum of three moves. */ - - if (! TARGET_SIMD && GET_MODE_SIZE (from) == 128 && GET_MODE_SIZE (to) == 128) - return regmove_cost->GP2FP + regmove_cost->FP2GP + regmove_cost->FP2FP; - - return regmove_cost->FP2FP; -} - -static int -aarch64_memory_move_cost (enum machine_mode mode ATTRIBUTE_UNUSED, - reg_class_t rclass ATTRIBUTE_UNUSED, - bool in ATTRIBUTE_UNUSED) -{ - return aarch64_tune_params->memmov_cost; -} - -static void initialize_aarch64_code_model (void); - -/* Parse the architecture extension string. */ - -static void -aarch64_parse_extension (char *str) -{ - /* The extension string is parsed left to right. */ - const struct aarch64_option_extension *opt = NULL; - - /* Flag to say whether we are adding or removing an extension. */ - int adding_ext = -1; - - while (str != NULL && *str != 0) - { - char *ext; - size_t len; - - str++; - ext = strchr (str, '+'); - - if (ext != NULL) - len = ext - str; - else - len = strlen (str); - - if (len >= 2 && strncmp (str, "no", 2) == 0) - { - adding_ext = 0; - len -= 2; - str += 2; - } - else if (len > 0) - adding_ext = 1; - - if (len == 0) - { - error ("missing feature modifier after %qs", "+no"); - return; - } - - /* Scan over the extensions table trying to find an exact match. */ - for (opt = all_extensions; opt->name != NULL; opt++) - { - if (strlen (opt->name) == len && strncmp (opt->name, str, len) == 0) - { - /* Add or remove the extension. */ - if (adding_ext) - aarch64_isa_flags |= opt->flags_on; - else - aarch64_isa_flags &= ~(opt->flags_off); - break; - } - } - - if (opt->name == NULL) - { - /* Extension not found in list. */ - error ("unknown feature modifier %qs", str); - return; - } - - str = ext; - }; - - return; -} - -/* Parse the ARCH string. */ - -static void -aarch64_parse_arch (void) -{ - char *ext; - const struct processor *arch; - char *str = (char *) alloca (strlen (aarch64_arch_string) + 1); - size_t len; - - strcpy (str, aarch64_arch_string); - - ext = strchr (str, '+'); - - if (ext != NULL) - len = ext - str; - else - len = strlen (str); - - if (len == 0) - { - error ("missing arch name in -march=%qs", str); - return; - } - - /* Loop through the list of supported ARCHs to find a match. */ - for (arch = all_architectures; arch->name != NULL; arch++) - { - if (strlen (arch->name) == len && strncmp (arch->name, str, len) == 0) - { - selected_arch = arch; - aarch64_isa_flags = selected_arch->flags; - selected_cpu = &all_cores[selected_arch->core]; - - if (ext != NULL) - { - /* ARCH string contains at least one extension. */ - aarch64_parse_extension (ext); - } - - return; - } - } - - /* ARCH name not found in list. */ - error ("unknown value %qs for -march", str); - return; -} - -/* Parse the CPU string. */ - -static void -aarch64_parse_cpu (void) -{ - char *ext; - const struct processor *cpu; - char *str = (char *) alloca (strlen (aarch64_cpu_string) + 1); - size_t len; - - strcpy (str, aarch64_cpu_string); - - ext = strchr (str, '+'); - - if (ext != NULL) - len = ext - str; - else - len = strlen (str); - - if (len == 0) - { - error ("missing cpu name in -mcpu=%qs", str); - return; - } - - /* Loop through the list of supported CPUs to find a match. */ - for (cpu = all_cores; cpu->name != NULL; cpu++) - { - if (strlen (cpu->name) == len && strncmp (cpu->name, str, len) == 0) - { - selected_cpu = cpu; - aarch64_isa_flags = selected_cpu->flags; - - if (ext != NULL) - { - /* CPU string contains at least one extension. */ - aarch64_parse_extension (ext); - } - - return; - } - } - - /* CPU name not found in list. */ - error ("unknown value %qs for -mcpu", str); - return; -} - -/* Parse the TUNE string. */ - -static void -aarch64_parse_tune (void) -{ - const struct processor *cpu; - char *str = (char *) alloca (strlen (aarch64_tune_string) + 1); - strcpy (str, aarch64_tune_string); - - /* Loop through the list of supported CPUs to find a match. */ - for (cpu = all_cores; cpu->name != NULL; cpu++) - { - if (strcmp (cpu->name, str) == 0) - { - selected_tune = cpu; - return; - } - } - - /* CPU name not found in list. */ - error ("unknown value %qs for -mtune", str); - return; -} - - -/* Implement TARGET_OPTION_OVERRIDE. */ - -static void -aarch64_override_options (void) -{ - /* march wins over mcpu, so when march is defined, mcpu takes the same value, - otherwise march remains undefined. mtune can be used with either march or - mcpu. */ - - if (aarch64_arch_string) - { - aarch64_parse_arch (); - aarch64_cpu_string = NULL; - } - - if (aarch64_cpu_string) - { - aarch64_parse_cpu (); - selected_arch = NULL; - } - - if (aarch64_tune_string) - { - aarch64_parse_tune (); - } - - initialize_aarch64_code_model (); - - aarch64_build_bitmask_table (); - - /* This target defaults to strict volatile bitfields. */ - if (flag_strict_volatile_bitfields < 0 && abi_version_at_least (2)) - flag_strict_volatile_bitfields = 1; - - /* If the user did not specify a processor, choose the default - one for them. This will be the CPU set during configuration using - --with-cpu, otherwise it is "generic". */ - if (!selected_cpu) - { - selected_cpu = &all_cores[TARGET_CPU_DEFAULT & 0x3f]; - aarch64_isa_flags = TARGET_CPU_DEFAULT >> 6; - } - - gcc_assert (selected_cpu); - - /* The selected cpu may be an architecture, so lookup tuning by core ID. */ - if (!selected_tune) - selected_tune = &all_cores[selected_cpu->core]; - - aarch64_tune_flags = selected_tune->flags; - aarch64_tune = selected_tune->core; - aarch64_tune_params = selected_tune->tune; - - aarch64_override_options_after_change (); -} - -/* Implement targetm.override_options_after_change. */ - -static void -aarch64_override_options_after_change (void) -{ - faked_omit_frame_pointer = false; - - /* To omit leaf frame pointers, we need to turn flag_omit_frame_pointer on so - that aarch64_frame_pointer_required will be called. We need to remember - whether flag_omit_frame_pointer was turned on normally or just faked. */ - - if (flag_omit_leaf_frame_pointer && !flag_omit_frame_pointer) - { - flag_omit_frame_pointer = true; - faked_omit_frame_pointer = true; - } -} - -static struct machine_function * -aarch64_init_machine_status (void) -{ - struct machine_function *machine; - machine = ggc_alloc_cleared_machine_function (); - return machine; -} - -void -aarch64_init_expanders (void) -{ - init_machine_status = aarch64_init_machine_status; -} - -/* A checking mechanism for the implementation of the various code models. */ -static void -initialize_aarch64_code_model (void) -{ - if (flag_pic) - { - switch (aarch64_cmodel_var) - { - case AARCH64_CMODEL_TINY: - aarch64_cmodel = AARCH64_CMODEL_TINY_PIC; - break; - case AARCH64_CMODEL_SMALL: - aarch64_cmodel = AARCH64_CMODEL_SMALL_PIC; - break; - case AARCH64_CMODEL_LARGE: - sorry ("code model %qs with -f%s", "large", - flag_pic > 1 ? "PIC" : "pic"); - default: - gcc_unreachable (); - } - } - else - aarch64_cmodel = aarch64_cmodel_var; -} - -/* Return true if SYMBOL_REF X binds locally. */ - -static bool -aarch64_symbol_binds_local_p (const_rtx x) -{ - return (SYMBOL_REF_DECL (x) - ? targetm.binds_local_p (SYMBOL_REF_DECL (x)) - : SYMBOL_REF_LOCAL_P (x)); -} - -/* Return true if SYMBOL_REF X is thread local */ -static bool -aarch64_tls_symbol_p (rtx x) -{ - if (! TARGET_HAVE_TLS) - return false; - - if (GET_CODE (x) != SYMBOL_REF) - return false; - - return SYMBOL_REF_TLS_MODEL (x) != 0; -} - -/* Classify a TLS symbol into one of the TLS kinds. */ -enum aarch64_symbol_type -aarch64_classify_tls_symbol (rtx x) -{ - enum tls_model tls_kind = tls_symbolic_operand_type (x); - - switch (tls_kind) - { - case TLS_MODEL_GLOBAL_DYNAMIC: - case TLS_MODEL_LOCAL_DYNAMIC: - return TARGET_TLS_DESC ? SYMBOL_SMALL_TLSDESC : SYMBOL_SMALL_TLSGD; - - case TLS_MODEL_INITIAL_EXEC: - return SYMBOL_SMALL_GOTTPREL; - - case TLS_MODEL_LOCAL_EXEC: - return SYMBOL_SMALL_TPREL; - - case TLS_MODEL_EMULATED: - case TLS_MODEL_NONE: - return SYMBOL_FORCE_TO_MEM; - - default: - gcc_unreachable (); - } -} - -/* Return the method that should be used to access SYMBOL_REF or - LABEL_REF X in context CONTEXT. */ -enum aarch64_symbol_type -aarch64_classify_symbol (rtx x, - enum aarch64_symbol_context context ATTRIBUTE_UNUSED) -{ - if (GET_CODE (x) == LABEL_REF) - { - switch (aarch64_cmodel) - { - case AARCH64_CMODEL_LARGE: - return SYMBOL_FORCE_TO_MEM; - - case AARCH64_CMODEL_TINY_PIC: - case AARCH64_CMODEL_TINY: - case AARCH64_CMODEL_SMALL_PIC: - case AARCH64_CMODEL_SMALL: - return SYMBOL_SMALL_ABSOLUTE; - - default: - gcc_unreachable (); - } - } - - gcc_assert (GET_CODE (x) == SYMBOL_REF); - - switch (aarch64_cmodel) - { - case AARCH64_CMODEL_LARGE: - return SYMBOL_FORCE_TO_MEM; - - case AARCH64_CMODEL_TINY: - case AARCH64_CMODEL_SMALL: - - /* This is needed to get DFmode, TImode constants to be loaded off - the constant pool. Is it necessary to dump TImode values into - the constant pool. We don't handle TImode constant loads properly - yet and hence need to use the constant pool. */ - if (CONSTANT_POOL_ADDRESS_P (x)) - return SYMBOL_FORCE_TO_MEM; - - if (aarch64_tls_symbol_p (x)) - return aarch64_classify_tls_symbol (x); - - if (SYMBOL_REF_WEAK (x)) - return SYMBOL_FORCE_TO_MEM; - - return SYMBOL_SMALL_ABSOLUTE; - - case AARCH64_CMODEL_TINY_PIC: - case AARCH64_CMODEL_SMALL_PIC: - - if (CONSTANT_POOL_ADDRESS_P (x)) - return SYMBOL_FORCE_TO_MEM; - - if (aarch64_tls_symbol_p (x)) - return aarch64_classify_tls_symbol (x); - - if (!aarch64_symbol_binds_local_p (x)) - return SYMBOL_SMALL_GOT; - - return SYMBOL_SMALL_ABSOLUTE; - - default: - gcc_unreachable (); - } - /* By default push everything into the constant pool. */ - return SYMBOL_FORCE_TO_MEM; -} - -/* Return true if X is a symbolic constant that can be used in context - CONTEXT. If it is, store the type of the symbol in *SYMBOL_TYPE. */ - -bool -aarch64_symbolic_constant_p (rtx x, enum aarch64_symbol_context context, - enum aarch64_symbol_type *symbol_type) -{ - rtx offset; - split_const (x, &x, &offset); - if (GET_CODE (x) == SYMBOL_REF || GET_CODE (x) == LABEL_REF) - *symbol_type = aarch64_classify_symbol (x, context); - else - return false; - - /* No checking of offset at this point. */ - return true; -} - -bool -aarch64_constant_address_p (rtx x) -{ - return (CONSTANT_P (x) && memory_address_p (DImode, x)); -} - -bool -aarch64_legitimate_pic_operand_p (rtx x) -{ - if (GET_CODE (x) == SYMBOL_REF - || (GET_CODE (x) == CONST - && GET_CODE (XEXP (x, 0)) == PLUS - && GET_CODE (XEXP (XEXP (x, 0), 0)) == SYMBOL_REF)) - return false; - - return true; -} - -/* Return true if X holds either a quarter-precision or - floating-point +0.0 constant. */ -static bool -aarch64_valid_floating_const (enum machine_mode mode, rtx x) -{ - if (!CONST_DOUBLE_P (x)) - return false; - - /* TODO: We could handle moving 0.0 to a TFmode register, - but first we would like to refactor the movtf_aarch64 - to be more amicable to split moves properly and - correctly gate on TARGET_SIMD. For now - reject all - constants which are not to SFmode or DFmode registers. */ - if (!(mode == SFmode || mode == DFmode)) - return false; - - if (aarch64_float_const_zero_rtx_p (x)) - return true; - return aarch64_float_const_representable_p (x); -} - -static bool -aarch64_legitimate_constant_p (enum machine_mode mode, rtx x) -{ - /* Do not allow vector struct mode constants. We could support - 0 and -1 easily, but they need support in aarch64-simd.md. */ - if (TARGET_SIMD && aarch64_vect_struct_mode_p (mode)) - return false; - - /* This could probably go away because - we now decompose CONST_INTs according to expand_mov_immediate. */ - if ((GET_CODE (x) == CONST_VECTOR - && aarch64_simd_valid_immediate (x, mode, false, - NULL, NULL, NULL, NULL, NULL) != -1) - || CONST_INT_P (x) || aarch64_valid_floating_const (mode, x)) - return !targetm.cannot_force_const_mem (mode, x); - - if (GET_CODE (x) == HIGH - && aarch64_valid_symref (XEXP (x, 0), GET_MODE (XEXP (x, 0)))) - return true; - - return aarch64_constant_address_p (x); -} - -rtx -aarch64_load_tp (rtx target) -{ - if (!target - || GET_MODE (target) != Pmode - || !register_operand (target, Pmode)) - target = gen_reg_rtx (Pmode); - - /* Can return in any reg. */ - emit_insn (gen_aarch64_load_tp_hard (target)); - return target; -} - -/* On AAPCS systems, this is the "struct __va_list". */ -static GTY(()) tree va_list_type; - -/* Implement TARGET_BUILD_BUILTIN_VA_LIST. - Return the type to use as __builtin_va_list. - - AAPCS64 \S 7.1.4 requires that va_list be a typedef for a type defined as: - - struct __va_list - { - void *__stack; - void *__gr_top; - void *__vr_top; - int __gr_offs; - int __vr_offs; - }; */ - -static tree -aarch64_build_builtin_va_list (void) -{ - tree va_list_name; - tree f_stack, f_grtop, f_vrtop, f_groff, f_vroff; - - /* Create the type. */ - va_list_type = lang_hooks.types.make_type (RECORD_TYPE); - /* Give it the required name. */ - va_list_name = build_decl (BUILTINS_LOCATION, - TYPE_DECL, - get_identifier ("__va_list"), - va_list_type); - DECL_ARTIFICIAL (va_list_name) = 1; - TYPE_NAME (va_list_type) = va_list_name; - TYPE_STUB_DECL (va_list_type) = va_list_name; - - /* Create the fields. */ - f_stack = build_decl (BUILTINS_LOCATION, - FIELD_DECL, get_identifier ("__stack"), - ptr_type_node); - f_grtop = build_decl (BUILTINS_LOCATION, - FIELD_DECL, get_identifier ("__gr_top"), - ptr_type_node); - f_vrtop = build_decl (BUILTINS_LOCATION, - FIELD_DECL, get_identifier ("__vr_top"), - ptr_type_node); - f_groff = build_decl (BUILTINS_LOCATION, - FIELD_DECL, get_identifier ("__gr_offs"), - integer_type_node); - f_vroff = build_decl (BUILTINS_LOCATION, - FIELD_DECL, get_identifier ("__vr_offs"), - integer_type_node); - - DECL_ARTIFICIAL (f_stack) = 1; - DECL_ARTIFICIAL (f_grtop) = 1; - DECL_ARTIFICIAL (f_vrtop) = 1; - DECL_ARTIFICIAL (f_groff) = 1; - DECL_ARTIFICIAL (f_vroff) = 1; - - DECL_FIELD_CONTEXT (f_stack) = va_list_type; - DECL_FIELD_CONTEXT (f_grtop) = va_list_type; - DECL_FIELD_CONTEXT (f_vrtop) = va_list_type; - DECL_FIELD_CONTEXT (f_groff) = va_list_type; - DECL_FIELD_CONTEXT (f_vroff) = va_list_type; - - TYPE_FIELDS (va_list_type) = f_stack; - DECL_CHAIN (f_stack) = f_grtop; - DECL_CHAIN (f_grtop) = f_vrtop; - DECL_CHAIN (f_vrtop) = f_groff; - DECL_CHAIN (f_groff) = f_vroff; - - /* Compute its layout. */ - layout_type (va_list_type); - - return va_list_type; -} - -/* Implement TARGET_EXPAND_BUILTIN_VA_START. */ -static void -aarch64_expand_builtin_va_start (tree valist, rtx nextarg ATTRIBUTE_UNUSED) -{ - const CUMULATIVE_ARGS *cum; - tree f_stack, f_grtop, f_vrtop, f_groff, f_vroff; - tree stack, grtop, vrtop, groff, vroff; - tree t; - int gr_save_area_size; - int vr_save_area_size; - int vr_offset; - - cum = &crtl->args.info; - gr_save_area_size - = (NUM_ARG_REGS - cum->aapcs_ncrn) * UNITS_PER_WORD; - vr_save_area_size - = (NUM_FP_ARG_REGS - cum->aapcs_nvrn) * UNITS_PER_VREG; - - if (TARGET_GENERAL_REGS_ONLY) - { - if (cum->aapcs_nvrn > 0) - sorry ("%qs and floating point or vector arguments", - "-mgeneral-regs-only"); - vr_save_area_size = 0; - } - - f_stack = TYPE_FIELDS (va_list_type_node); - f_grtop = DECL_CHAIN (f_stack); - f_vrtop = DECL_CHAIN (f_grtop); - f_groff = DECL_CHAIN (f_vrtop); - f_vroff = DECL_CHAIN (f_groff); - - stack = build3 (COMPONENT_REF, TREE_TYPE (f_stack), valist, f_stack, - NULL_TREE); - grtop = build3 (COMPONENT_REF, TREE_TYPE (f_grtop), valist, f_grtop, - NULL_TREE); - vrtop = build3 (COMPONENT_REF, TREE_TYPE (f_vrtop), valist, f_vrtop, - NULL_TREE); - groff = build3 (COMPONENT_REF, TREE_TYPE (f_groff), valist, f_groff, - NULL_TREE); - vroff = build3 (COMPONENT_REF, TREE_TYPE (f_vroff), valist, f_vroff, - NULL_TREE); - - /* Emit code to initialize STACK, which points to the next varargs stack - argument. CUM->AAPCS_STACK_SIZE gives the number of stack words used - by named arguments. STACK is 8-byte aligned. */ - t = make_tree (TREE_TYPE (stack), virtual_incoming_args_rtx); - if (cum->aapcs_stack_size > 0) - t = fold_build_pointer_plus_hwi (t, cum->aapcs_stack_size * UNITS_PER_WORD); - t = build2 (MODIFY_EXPR, TREE_TYPE (stack), stack, t); - expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); - - /* Emit code to initialize GRTOP, the top of the GR save area. - virtual_incoming_args_rtx should have been 16 byte aligned. */ - t = make_tree (TREE_TYPE (grtop), virtual_incoming_args_rtx); - t = build2 (MODIFY_EXPR, TREE_TYPE (grtop), grtop, t); - expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); - - /* Emit code to initialize VRTOP, the top of the VR save area. - This address is gr_save_area_bytes below GRTOP, rounded - down to the next 16-byte boundary. */ - t = make_tree (TREE_TYPE (vrtop), virtual_incoming_args_rtx); - vr_offset = AARCH64_ROUND_UP (gr_save_area_size, - STACK_BOUNDARY / BITS_PER_UNIT); - - if (vr_offset) - t = fold_build_pointer_plus_hwi (t, -vr_offset); - t = build2 (MODIFY_EXPR, TREE_TYPE (vrtop), vrtop, t); - expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); - - /* Emit code to initialize GROFF, the offset from GRTOP of the - next GPR argument. */ - t = build2 (MODIFY_EXPR, TREE_TYPE (groff), groff, - build_int_cst (TREE_TYPE (groff), -gr_save_area_size)); - expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); - - /* Likewise emit code to initialize VROFF, the offset from FTOP - of the next VR argument. */ - t = build2 (MODIFY_EXPR, TREE_TYPE (vroff), vroff, - build_int_cst (TREE_TYPE (vroff), -vr_save_area_size)); - expand_expr (t, const0_rtx, VOIDmode, EXPAND_NORMAL); -} - -/* Implement TARGET_GIMPLIFY_VA_ARG_EXPR. */ - -static tree -aarch64_gimplify_va_arg_expr (tree valist, tree type, gimple_seq *pre_p, - gimple_seq *post_p ATTRIBUTE_UNUSED) -{ - tree addr; - bool indirect_p; - bool is_ha; /* is HFA or HVA. */ - bool dw_align; /* double-word align. */ - enum machine_mode ag_mode = VOIDmode; - int nregs; - enum machine_mode mode; - - tree f_stack, f_grtop, f_vrtop, f_groff, f_vroff; - tree stack, f_top, f_off, off, arg, roundup, on_stack; - HOST_WIDE_INT size, rsize, adjust, align; - tree t, u, cond1, cond2; - - indirect_p = pass_by_reference (NULL, TYPE_MODE (type), type, false); - if (indirect_p) - type = build_pointer_type (type); - - mode = TYPE_MODE (type); - - f_stack = TYPE_FIELDS (va_list_type_node); - f_grtop = DECL_CHAIN (f_stack); - f_vrtop = DECL_CHAIN (f_grtop); - f_groff = DECL_CHAIN (f_vrtop); - f_vroff = DECL_CHAIN (f_groff); - - stack = build3 (COMPONENT_REF, TREE_TYPE (f_stack), unshare_expr (valist), - f_stack, NULL_TREE); - size = int_size_in_bytes (type); - align = aarch64_function_arg_alignment (mode, type) / BITS_PER_UNIT; - - dw_align = false; - adjust = 0; - if (aarch64_vfp_is_call_or_return_candidate (mode, - type, - &ag_mode, - &nregs, - &is_ha)) - { - /* TYPE passed in fp/simd registers. */ - if (TARGET_GENERAL_REGS_ONLY) - sorry ("%qs and floating point or vector arguments", - "-mgeneral-regs-only"); - - f_top = build3 (COMPONENT_REF, TREE_TYPE (f_vrtop), - unshare_expr (valist), f_vrtop, NULL_TREE); - f_off = build3 (COMPONENT_REF, TREE_TYPE (f_vroff), - unshare_expr (valist), f_vroff, NULL_TREE); - - rsize = nregs * UNITS_PER_VREG; - - if (is_ha) - { - if (BYTES_BIG_ENDIAN && GET_MODE_SIZE (ag_mode) < UNITS_PER_VREG) - adjust = UNITS_PER_VREG - GET_MODE_SIZE (ag_mode); - } - else if (BLOCK_REG_PADDING (mode, type, 1) == downward - && size < UNITS_PER_VREG) - { - adjust = UNITS_PER_VREG - size; - } - } - else - { - /* TYPE passed in general registers. */ - f_top = build3 (COMPONENT_REF, TREE_TYPE (f_grtop), - unshare_expr (valist), f_grtop, NULL_TREE); - f_off = build3 (COMPONENT_REF, TREE_TYPE (f_groff), - unshare_expr (valist), f_groff, NULL_TREE); - rsize = (size + UNITS_PER_WORD - 1) & -UNITS_PER_WORD; - nregs = rsize / UNITS_PER_WORD; - - if (align > 8) - dw_align = true; - - if (BLOCK_REG_PADDING (mode, type, 1) == downward - && size < UNITS_PER_WORD) - { - adjust = UNITS_PER_WORD - size; - } - } - - /* Get a local temporary for the field value. */ - off = get_initialized_tmp_var (f_off, pre_p, NULL); - - /* Emit code to branch if off >= 0. */ - t = build2 (GE_EXPR, boolean_type_node, off, - build_int_cst (TREE_TYPE (off), 0)); - cond1 = build3 (COND_EXPR, ptr_type_node, t, NULL_TREE, NULL_TREE); - - if (dw_align) - { - /* Emit: offs = (offs + 15) & -16. */ - t = build2 (PLUS_EXPR, TREE_TYPE (off), off, - build_int_cst (TREE_TYPE (off), 15)); - t = build2 (BIT_AND_EXPR, TREE_TYPE (off), t, - build_int_cst (TREE_TYPE (off), -16)); - roundup = build2 (MODIFY_EXPR, TREE_TYPE (off), off, t); - } - else - roundup = NULL; - - /* Update ap.__[g|v]r_offs */ - t = build2 (PLUS_EXPR, TREE_TYPE (off), off, - build_int_cst (TREE_TYPE (off), rsize)); - t = build2 (MODIFY_EXPR, TREE_TYPE (f_off), unshare_expr (f_off), t); - - /* String up. */ - if (roundup) - t = build2 (COMPOUND_EXPR, TREE_TYPE (t), roundup, t); - - /* [cond2] if (ap.__[g|v]r_offs > 0) */ - u = build2 (GT_EXPR, boolean_type_node, unshare_expr (f_off), - build_int_cst (TREE_TYPE (f_off), 0)); - cond2 = build3 (COND_EXPR, ptr_type_node, u, NULL_TREE, NULL_TREE); - - /* String up: make sure the assignment happens before the use. */ - t = build2 (COMPOUND_EXPR, TREE_TYPE (cond2), t, cond2); - COND_EXPR_ELSE (cond1) = t; - - /* Prepare the trees handling the argument that is passed on the stack; - the top level node will store in ON_STACK. */ - arg = get_initialized_tmp_var (stack, pre_p, NULL); - if (align > 8) - { - /* if (alignof(type) > 8) (arg = arg + 15) & -16; */ - t = fold_convert (intDI_type_node, arg); - t = build2 (PLUS_EXPR, TREE_TYPE (t), t, - build_int_cst (TREE_TYPE (t), 15)); - t = build2 (BIT_AND_EXPR, TREE_TYPE (t), t, - build_int_cst (TREE_TYPE (t), -16)); - t = fold_convert (TREE_TYPE (arg), t); - roundup = build2 (MODIFY_EXPR, TREE_TYPE (arg), arg, t); - } - else - roundup = NULL; - /* Advance ap.__stack */ - t = fold_convert (intDI_type_node, arg); - t = build2 (PLUS_EXPR, TREE_TYPE (t), t, - build_int_cst (TREE_TYPE (t), size + 7)); - t = build2 (BIT_AND_EXPR, TREE_TYPE (t), t, - build_int_cst (TREE_TYPE (t), -8)); - t = fold_convert (TREE_TYPE (arg), t); - t = build2 (MODIFY_EXPR, TREE_TYPE (stack), unshare_expr (stack), t); - /* String up roundup and advance. */ - if (roundup) - t = build2 (COMPOUND_EXPR, TREE_TYPE (t), roundup, t); - /* String up with arg */ - on_stack = build2 (COMPOUND_EXPR, TREE_TYPE (arg), t, arg); - /* Big-endianness related address adjustment. */ - if (BLOCK_REG_PADDING (mode, type, 1) == downward - && size < UNITS_PER_WORD) - { - t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (arg), arg, - size_int (UNITS_PER_WORD - size)); - on_stack = build2 (COMPOUND_EXPR, TREE_TYPE (arg), on_stack, t); - } - - COND_EXPR_THEN (cond1) = unshare_expr (on_stack); - COND_EXPR_THEN (cond2) = unshare_expr (on_stack); - - /* Adjustment to OFFSET in the case of BIG_ENDIAN. */ - t = off; - if (adjust) - t = build2 (PREINCREMENT_EXPR, TREE_TYPE (off), off, - build_int_cst (TREE_TYPE (off), adjust)); - - t = fold_convert (sizetype, t); - t = build2 (POINTER_PLUS_EXPR, TREE_TYPE (f_top), f_top, t); - - if (is_ha) - { - /* type ha; // treat as "struct {ftype field[n];}" - ... [computing offs] - for (i = 0; i <nregs; ++i, offs += 16) - ha.field[i] = *((ftype *)(ap.__vr_top + offs)); - return ha; */ - int i; - tree tmp_ha, field_t, field_ptr_t; - - /* Declare a local variable. */ - tmp_ha = create_tmp_var_raw (type, "ha"); - gimple_add_tmp_var (tmp_ha); - - /* Establish the base type. */ - switch (ag_mode) - { - case SFmode: - field_t = float_type_node; - field_ptr_t = float_ptr_type_node; - break; - case DFmode: - field_t = double_type_node; - field_ptr_t = double_ptr_type_node; - break; - case TFmode: - field_t = long_double_type_node; - field_ptr_t = long_double_ptr_type_node; - break; -/* The half precision and quad precision are not fully supported yet. Enable - the following code after the support is complete. Need to find the correct - type node for __fp16 *. */ -#if 0 - case HFmode: - field_t = float_type_node; - field_ptr_t = float_ptr_type_node; - break; -#endif - case V2SImode: - case V4SImode: - { - tree innertype = make_signed_type (GET_MODE_PRECISION (SImode)); - field_t = build_vector_type_for_mode (innertype, ag_mode); - field_ptr_t = build_pointer_type (field_t); - } - break; - default: - gcc_assert (0); - } - - /* *(field_ptr_t)&ha = *((field_ptr_t)vr_saved_area */ - tmp_ha = build1 (ADDR_EXPR, field_ptr_t, tmp_ha); - addr = t; - t = fold_convert (field_ptr_t, addr); - t = build2 (MODIFY_EXPR, field_t, - build1 (INDIRECT_REF, field_t, tmp_ha), - build1 (INDIRECT_REF, field_t, t)); - - /* ha.field[i] = *((field_ptr_t)vr_saved_area + i) */ - for (i = 1; i < nregs; ++i) - { - addr = fold_build_pointer_plus_hwi (addr, UNITS_PER_VREG); - u = fold_convert (field_ptr_t, addr); - u = build2 (MODIFY_EXPR, field_t, - build2 (MEM_REF, field_t, tmp_ha, - build_int_cst (field_ptr_t, - (i * - int_size_in_bytes (field_t)))), - build1 (INDIRECT_REF, field_t, u)); - t = build2 (COMPOUND_EXPR, TREE_TYPE (t), t, u); - } - - u = fold_convert (TREE_TYPE (f_top), tmp_ha); - t = build2 (COMPOUND_EXPR, TREE_TYPE (f_top), t, u); - } - - COND_EXPR_ELSE (cond2) = t; - addr = fold_convert (build_pointer_type (type), cond1); - addr = build_va_arg_indirect_ref (addr); - - if (indirect_p) - addr = build_va_arg_indirect_ref (addr); - - return addr; -} - -/* Implement TARGET_SETUP_INCOMING_VARARGS. */ - -static void -aarch64_setup_incoming_varargs (cumulative_args_t cum_v, enum machine_mode mode, - tree type, int *pretend_size ATTRIBUTE_UNUSED, - int no_rtl) -{ - CUMULATIVE_ARGS *cum = get_cumulative_args (cum_v); - CUMULATIVE_ARGS local_cum; - int gr_saved, vr_saved; - - /* The caller has advanced CUM up to, but not beyond, the last named - argument. Advance a local copy of CUM past the last "real" named - argument, to find out how many registers are left over. */ - local_cum = *cum; - aarch64_function_arg_advance (pack_cumulative_args(&local_cum), mode, type, true); - - /* Found out how many registers we need to save. */ - gr_saved = NUM_ARG_REGS - local_cum.aapcs_ncrn; - vr_saved = NUM_FP_ARG_REGS - local_cum.aapcs_nvrn; - - if (TARGET_GENERAL_REGS_ONLY) - { - if (local_cum.aapcs_nvrn > 0) - sorry ("%qs and floating point or vector arguments", - "-mgeneral-regs-only"); - vr_saved = 0; - } - - if (!no_rtl) - { - if (gr_saved > 0) - { - rtx ptr, mem; - - /* virtual_incoming_args_rtx should have been 16-byte aligned. */ - ptr = plus_constant (Pmode, virtual_incoming_args_rtx, - - gr_saved * UNITS_PER_WORD); - mem = gen_frame_mem (BLKmode, ptr); - set_mem_alias_set (mem, get_varargs_alias_set ()); - - move_block_from_reg (local_cum.aapcs_ncrn + R0_REGNUM, - mem, gr_saved); - } - if (vr_saved > 0) - { - /* We can't use move_block_from_reg, because it will use - the wrong mode, storing D regs only. */ - enum machine_mode mode = TImode; - int off, i; - - /* Set OFF to the offset from virtual_incoming_args_rtx of - the first vector register. The VR save area lies below - the GR one, and is aligned to 16 bytes. */ - off = -AARCH64_ROUND_UP (gr_saved * UNITS_PER_WORD, - STACK_BOUNDARY / BITS_PER_UNIT); - off -= vr_saved * UNITS_PER_VREG; - - for (i = local_cum.aapcs_nvrn; i < NUM_FP_ARG_REGS; ++i) - { - rtx ptr, mem; - - ptr = plus_constant (Pmode, virtual_incoming_args_rtx, off); - mem = gen_frame_mem (mode, ptr); - set_mem_alias_set (mem, get_varargs_alias_set ()); - aarch64_emit_move (mem, gen_rtx_REG (mode, V0_REGNUM + i)); - off += UNITS_PER_VREG; - } - } - } - - /* We don't save the size into *PRETEND_SIZE because we want to avoid - any complication of having crtl->args.pretend_args_size changed. */ - cfun->machine->saved_varargs_size - = (AARCH64_ROUND_UP (gr_saved * UNITS_PER_WORD, - STACK_BOUNDARY / BITS_PER_UNIT) - + vr_saved * UNITS_PER_VREG); -} - -static void -aarch64_conditional_register_usage (void) -{ - int i; - if (!TARGET_FLOAT) - { - for (i = V0_REGNUM; i <= V31_REGNUM; i++) - { - fixed_regs[i] = 1; - call_used_regs[i] = 1; - } - } -} - -/* Walk down the type tree of TYPE counting consecutive base elements. - If *MODEP is VOIDmode, then set it to the first valid floating point - type. If a non-floating point type is found, or if a floating point - type that doesn't match a non-VOIDmode *MODEP is found, then return -1, - otherwise return the count in the sub-tree. */ -static int -aapcs_vfp_sub_candidate (const_tree type, enum machine_mode *modep) -{ - enum machine_mode mode; - HOST_WIDE_INT size; - - switch (TREE_CODE (type)) - { - case REAL_TYPE: - mode = TYPE_MODE (type); - if (mode != DFmode && mode != SFmode && mode != TFmode) - return -1; - - if (*modep == VOIDmode) - *modep = mode; - - if (*modep == mode) - return 1; - - break; - - case COMPLEX_TYPE: - mode = TYPE_MODE (TREE_TYPE (type)); - if (mode != DFmode && mode != SFmode && mode != TFmode) - return -1; - - if (*modep == VOIDmode) - *modep = mode; - - if (*modep == mode) - return 2; - - break; - - case VECTOR_TYPE: - /* Use V2SImode and V4SImode as representatives of all 64-bit - and 128-bit vector types. */ - size = int_size_in_bytes (type); - switch (size) - { - case 8: - mode = V2SImode; - break; - case 16: - mode = V4SImode; - break; - default: - return -1; - } - - if (*modep == VOIDmode) - *modep = mode; - - /* Vector modes are considered to be opaque: two vectors are - equivalent for the purposes of being homogeneous aggregates - if they are the same size. */ - if (*modep == mode) - return 1; - - break; - - case ARRAY_TYPE: - { - int count; - tree index = TYPE_DOMAIN (type); - - /* Can't handle incomplete types. */ - if (!COMPLETE_TYPE_P (type)) - return -1; - - count = aapcs_vfp_sub_candidate (TREE_TYPE (type), modep); - if (count == -1 - || !index - || !TYPE_MAX_VALUE (index) - || !host_integerp (TYPE_MAX_VALUE (index), 1) - || !TYPE_MIN_VALUE (index) - || !host_integerp (TYPE_MIN_VALUE (index), 1) - || count < 0) - return -1; - - count *= (1 + tree_low_cst (TYPE_MAX_VALUE (index), 1) - - tree_low_cst (TYPE_MIN_VALUE (index), 1)); - - /* There must be no padding. */ - if (!host_integerp (TYPE_SIZE (type), 1) - || (tree_low_cst (TYPE_SIZE (type), 1) - != count * GET_MODE_BITSIZE (*modep))) - return -1; - - return count; - } - - case RECORD_TYPE: - { - int count = 0; - int sub_count; - tree field; - - /* Can't handle incomplete types. */ - if (!COMPLETE_TYPE_P (type)) - return -1; - - for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field)) - { - if (TREE_CODE (field) != FIELD_DECL) - continue; - - sub_count = aapcs_vfp_sub_candidate (TREE_TYPE (field), modep); - if (sub_count < 0) - return -1; - count += sub_count; - } - - /* There must be no padding. */ - if (!host_integerp (TYPE_SIZE (type), 1) - || (tree_low_cst (TYPE_SIZE (type), 1) - != count * GET_MODE_BITSIZE (*modep))) - return -1; - - return count; - } - - case UNION_TYPE: - case QUAL_UNION_TYPE: - { - /* These aren't very interesting except in a degenerate case. */ - int count = 0; - int sub_count; - tree field; - - /* Can't handle incomplete types. */ - if (!COMPLETE_TYPE_P (type)) - return -1; - - for (field = TYPE_FIELDS (type); field; field = TREE_CHAIN (field)) - { - if (TREE_CODE (field) != FIELD_DECL) - continue; - - sub_count = aapcs_vfp_sub_candidate (TREE_TYPE (field), modep); - if (sub_count < 0) - return -1; - count = count > sub_count ? count : sub_count; - } - - /* There must be no padding. */ - if (!host_integerp (TYPE_SIZE (type), 1) - || (tree_low_cst (TYPE_SIZE (type), 1) - != count * GET_MODE_BITSIZE (*modep))) - return -1; - - return count; - } - - default: - break; - } - - return -1; -} - -/* Return TRUE if the type, as described by TYPE and MODE, is a composite - type as described in AAPCS64 \S 4.3. This includes aggregate, union and - array types. The C99 floating-point complex types are also considered - as composite types, according to AAPCS64 \S 7.1.1. The complex integer - types, which are GCC extensions and out of the scope of AAPCS64, are - treated as composite types here as well. - - Note that MODE itself is not sufficient in determining whether a type - is such a composite type or not. This is because - stor-layout.c:compute_record_mode may have already changed the MODE - (BLKmode) of a RECORD_TYPE TYPE to some other mode. For example, a - structure with only one field may have its MODE set to the mode of the - field. Also an integer mode whose size matches the size of the - RECORD_TYPE type may be used to substitute the original mode - (i.e. BLKmode) in certain circumstances. In other words, MODE cannot be - solely relied on. */ - -static bool -aarch64_composite_type_p (const_tree type, - enum machine_mode mode) -{ - if (type && (AGGREGATE_TYPE_P (type) || TREE_CODE (type) == COMPLEX_TYPE)) - return true; - - if (mode == BLKmode - || GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT - || GET_MODE_CLASS (mode) == MODE_COMPLEX_INT) - return true; - - return false; -} - -/* Return TRUE if the type, as described by TYPE and MODE, is a short vector - type as described in AAPCS64 \S 4.1.2. - - See the comment above aarch64_composite_type_p for the notes on MODE. */ - -static bool -aarch64_short_vector_p (const_tree type, - enum machine_mode mode) -{ - HOST_WIDE_INT size = -1; - - if (type && TREE_CODE (type) == VECTOR_TYPE) - size = int_size_in_bytes (type); - else if (!aarch64_composite_type_p (type, mode) - && (GET_MODE_CLASS (mode) == MODE_VECTOR_INT - || GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT)) - size = GET_MODE_SIZE (mode); - - return (size == 8 || size == 16) ? true : false; -} - -/* Return TRUE if an argument, whose type is described by TYPE and MODE, - shall be passed or returned in simd/fp register(s) (providing these - parameter passing registers are available). - - Upon successful return, *COUNT returns the number of needed registers, - *BASE_MODE returns the mode of the individual register and when IS_HAF - is not NULL, *IS_HA indicates whether or not the argument is a homogeneous - floating-point aggregate or a homogeneous short-vector aggregate. */ - -static bool -aarch64_vfp_is_call_or_return_candidate (enum machine_mode mode, - const_tree type, - enum machine_mode *base_mode, - int *count, - bool *is_ha) -{ - enum machine_mode new_mode = VOIDmode; - bool composite_p = aarch64_composite_type_p (type, mode); - - if (is_ha != NULL) *is_ha = false; - - if ((!composite_p && GET_MODE_CLASS (mode) == MODE_FLOAT) - || aarch64_short_vector_p (type, mode)) - { - *count = 1; - new_mode = mode; - } - else if (GET_MODE_CLASS (mode) == MODE_COMPLEX_FLOAT) - { - if (is_ha != NULL) *is_ha = true; - *count = 2; - new_mode = GET_MODE_INNER (mode); - } - else if (type && composite_p) - { - int ag_count = aapcs_vfp_sub_candidate (type, &new_mode); - - if (ag_count > 0 && ag_count <= HA_MAX_NUM_FLDS) - { - if (is_ha != NULL) *is_ha = true; - *count = ag_count; - } - else - return false; - } - else - return false; - - *base_mode = new_mode; - return true; -} - -/* Implement TARGET_STRUCT_VALUE_RTX. */ - -static rtx -aarch64_struct_value_rtx (tree fndecl ATTRIBUTE_UNUSED, - int incoming ATTRIBUTE_UNUSED) -{ - return gen_rtx_REG (Pmode, AARCH64_STRUCT_VALUE_REGNUM); -} - -/* Implements target hook vector_mode_supported_p. */ -static bool -aarch64_vector_mode_supported_p (enum machine_mode mode) -{ - if (TARGET_SIMD - && (mode == V4SImode || mode == V8HImode - || mode == V16QImode || mode == V2DImode - || mode == V2SImode || mode == V4HImode - || mode == V8QImode || mode == V2SFmode - || mode == V4SFmode || mode == V2DFmode)) - return true; - - return false; -} - -/* Return quad mode as the preferred SIMD mode. */ -static enum machine_mode -aarch64_preferred_simd_mode (enum machine_mode mode) -{ - if (TARGET_SIMD) - switch (mode) - { - case DFmode: - return V2DFmode; - case SFmode: - return V4SFmode; - case SImode: - return V4SImode; - case HImode: - return V8HImode; - case QImode: - return V16QImode; - case DImode: - return V2DImode; - break; - - default:; - } - return word_mode; -} - -/* Return the bitmask of possible vector sizes for the vectorizer - to iterate over. */ -static unsigned int -aarch64_autovectorize_vector_sizes (void) -{ - return (16 | 8); -} - -/* A table to help perform AArch64-specific name mangling for AdvSIMD - vector types in order to conform to the AAPCS64 (see "Procedure - Call Standard for the ARM 64-bit Architecture", Appendix A). To - qualify for emission with the mangled names defined in that document, - a vector type must not only be of the correct mode but also be - composed of AdvSIMD vector element types (e.g. - _builtin_aarch64_simd_qi); these types are registered by - aarch64_init_simd_builtins (). In other words, vector types defined - in other ways e.g. via vector_size attribute will get default - mangled names. */ -typedef struct -{ - enum machine_mode mode; - const char *element_type_name; - const char *mangled_name; -} aarch64_simd_mangle_map_entry; - -static aarch64_simd_mangle_map_entry aarch64_simd_mangle_map[] = { - /* 64-bit containerized types. */ - { V8QImode, "__builtin_aarch64_simd_qi", "10__Int8x8_t" }, - { V8QImode, "__builtin_aarch64_simd_uqi", "11__Uint8x8_t" }, - { V4HImode, "__builtin_aarch64_simd_hi", "11__Int16x4_t" }, - { V4HImode, "__builtin_aarch64_simd_uhi", "12__Uint16x4_t" }, - { V2SImode, "__builtin_aarch64_simd_si", "11__Int32x2_t" }, - { V2SImode, "__builtin_aarch64_simd_usi", "12__Uint32x2_t" }, - { V2SFmode, "__builtin_aarch64_simd_sf", "13__Float32x2_t" }, - { V8QImode, "__builtin_aarch64_simd_poly8", "11__Poly8x8_t" }, - { V4HImode, "__builtin_aarch64_simd_poly16", "12__Poly16x4_t" }, - /* 128-bit containerized types. */ - { V16QImode, "__builtin_aarch64_simd_qi", "11__Int8x16_t" }, - { V16QImode, "__builtin_aarch64_simd_uqi", "12__Uint8x16_t" }, - { V8HImode, "__builtin_aarch64_simd_hi", "11__Int16x8_t" }, - { V8HImode, "__builtin_aarch64_simd_uhi", "12__Uint16x8_t" }, - { V4SImode, "__builtin_aarch64_simd_si", "11__Int32x4_t" }, - { V4SImode, "__builtin_aarch64_simd_usi", "12__Uint32x4_t" }, - { V2DImode, "__builtin_aarch64_simd_di", "11__Int64x2_t" }, - { V2DImode, "__builtin_aarch64_simd_udi", "12__Uint64x2_t" }, - { V4SFmode, "__builtin_aarch64_simd_sf", "13__Float32x4_t" }, - { V2DFmode, "__builtin_aarch64_simd_df", "13__Float64x2_t" }, - { V16QImode, "__builtin_aarch64_simd_poly8", "12__Poly8x16_t" }, - { V8HImode, "__builtin_aarch64_simd_poly16", "12__Poly16x8_t" }, - { VOIDmode, NULL, NULL } -}; - -/* Implement TARGET_MANGLE_TYPE. */ - -static const char * -aarch64_mangle_type (const_tree type) -{ - /* The AArch64 ABI documents say that "__va_list" has to be - managled as if it is in the "std" namespace. */ - if (lang_hooks.types_compatible_p (CONST_CAST_TREE (type), va_list_type)) - return "St9__va_list"; - - /* Check the mode of the vector type, and the name of the vector - element type, against the table. */ - if (TREE_CODE (type) == VECTOR_TYPE) - { - aarch64_simd_mangle_map_entry *pos = aarch64_simd_mangle_map; - - while (pos->mode != VOIDmode) - { - tree elt_type = TREE_TYPE (type); - - if (pos->mode == TYPE_MODE (type) - && TREE_CODE (TYPE_NAME (elt_type)) == TYPE_DECL - && !strcmp (IDENTIFIER_POINTER (DECL_NAME (TYPE_NAME (elt_type))), - pos->element_type_name)) - return pos->mangled_name; - - pos++; - } - } - - /* Use the default mangling. */ - return NULL; -} - -/* Return the equivalent letter for size. */ -static unsigned char -sizetochar (int size) -{ - switch (size) - { - case 64: return 'd'; - case 32: return 's'; - case 16: return 'h'; - case 8 : return 'b'; - default: gcc_unreachable (); - } -} - -/* Return true iff x is a uniform vector of floating-point - constants, and the constant can be represented in - quarter-precision form. Note, as aarch64_float_const_representable - rejects both +0.0 and -0.0, we will also reject +0.0 and -0.0. */ -static bool -aarch64_vect_float_const_representable_p (rtx x) -{ - int i = 0; - REAL_VALUE_TYPE r0, ri; - rtx x0, xi; - - if (GET_MODE_CLASS (GET_MODE (x)) != MODE_VECTOR_FLOAT) - return false; - - x0 = CONST_VECTOR_ELT (x, 0); - if (!CONST_DOUBLE_P (x0)) - return false; - - REAL_VALUE_FROM_CONST_DOUBLE (r0, x0); - - for (i = 1; i < CONST_VECTOR_NUNITS (x); i++) - { - xi = CONST_VECTOR_ELT (x, i); - if (!CONST_DOUBLE_P (xi)) - return false; - - REAL_VALUE_FROM_CONST_DOUBLE (ri, xi); - if (!REAL_VALUES_EQUAL (r0, ri)) - return false; - } - - return aarch64_float_const_representable_p (x0); -} - -/* TODO: This function returns values similar to those - returned by neon_valid_immediate in gcc/config/arm/arm.c - but the API here is different enough that these magic numbers - are not used. It should be sufficient to return true or false. */ -static int -aarch64_simd_valid_immediate (rtx op, enum machine_mode mode, int inverse, - rtx *modconst, int *elementwidth, - unsigned char *elementchar, - int *mvn, int *shift) -{ -#define CHECK(STRIDE, ELSIZE, CLASS, TEST, SHIFT, NEG) \ - matches = 1; \ - for (i = 0; i < idx; i += (STRIDE)) \ - if (!(TEST)) \ - matches = 0; \ - if (matches) \ - { \ - immtype = (CLASS); \ - elsize = (ELSIZE); \ - elchar = sizetochar (elsize); \ - eshift = (SHIFT); \ - emvn = (NEG); \ - break; \ - } - - unsigned int i, elsize = 0, idx = 0, n_elts = CONST_VECTOR_NUNITS (op); - unsigned int innersize = GET_MODE_SIZE (GET_MODE_INNER (mode)); - unsigned char bytes[16]; - unsigned char elchar = 0; - int immtype = -1, matches; - unsigned int invmask = inverse ? 0xff : 0; - int eshift, emvn; - - if (GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT) - { - bool simd_imm_zero = aarch64_simd_imm_zero_p (op, mode); - int elem_width = GET_MODE_BITSIZE (GET_MODE (CONST_VECTOR_ELT (op, 0))); - - if (!(simd_imm_zero - || aarch64_vect_float_const_representable_p (op))) - return -1; - - if (modconst) - *modconst = CONST_VECTOR_ELT (op, 0); - - if (elementwidth) - *elementwidth = elem_width; - - if (elementchar) - *elementchar = sizetochar (elem_width); - - if (shift) - *shift = 0; - - if (simd_imm_zero) - return 19; - else - return 18; - } - - /* Splat vector constant out into a byte vector. */ - for (i = 0; i < n_elts; i++) - { - rtx el = CONST_VECTOR_ELT (op, i); - unsigned HOST_WIDE_INT elpart; - unsigned int part, parts; - - if (GET_CODE (el) == CONST_INT) - { - elpart = INTVAL (el); - parts = 1; - } - else if (GET_CODE (el) == CONST_DOUBLE) - { - elpart = CONST_DOUBLE_LOW (el); - parts = 2; - } - else - gcc_unreachable (); - - for (part = 0; part < parts; part++) - { - unsigned int byte; - for (byte = 0; byte < innersize; byte++) - { - bytes[idx++] = (elpart & 0xff) ^ invmask; - elpart >>= BITS_PER_UNIT; - } - if (GET_CODE (el) == CONST_DOUBLE) - elpart = CONST_DOUBLE_HIGH (el); - } - } - - /* Sanity check. */ - gcc_assert (idx == GET_MODE_SIZE (mode)); - - do - { - CHECK (4, 32, 0, bytes[i] == bytes[0] && bytes[i + 1] == 0 - && bytes[i + 2] == 0 && bytes[i + 3] == 0, 0, 0); - - CHECK (4, 32, 1, bytes[i] == 0 && bytes[i + 1] == bytes[1] - && bytes[i + 2] == 0 && bytes[i + 3] == 0, 8, 0); - - CHECK (4, 32, 2, bytes[i] == 0 && bytes[i + 1] == 0 - && bytes[i + 2] == bytes[2] && bytes[i + 3] == 0, 16, 0); - - CHECK (4, 32, 3, bytes[i] == 0 && bytes[i + 1] == 0 - && bytes[i + 2] == 0 && bytes[i + 3] == bytes[3], 24, 0); - - CHECK (2, 16, 4, bytes[i] == bytes[0] && bytes[i + 1] == 0, 0, 0); - - CHECK (2, 16, 5, bytes[i] == 0 && bytes[i + 1] == bytes[1], 8, 0); - - CHECK (4, 32, 6, bytes[i] == bytes[0] && bytes[i + 1] == 0xff - && bytes[i + 2] == 0xff && bytes[i + 3] == 0xff, 0, 1); - - CHECK (4, 32, 7, bytes[i] == 0xff && bytes[i + 1] == bytes[1] - && bytes[i + 2] == 0xff && bytes[i + 3] == 0xff, 8, 1); - - CHECK (4, 32, 8, bytes[i] == 0xff && bytes[i + 1] == 0xff - && bytes[i + 2] == bytes[2] && bytes[i + 3] == 0xff, 16, 1); - - CHECK (4, 32, 9, bytes[i] == 0xff && bytes[i + 1] == 0xff - && bytes[i + 2] == 0xff && bytes[i + 3] == bytes[3], 24, 1); - - CHECK (2, 16, 10, bytes[i] == bytes[0] && bytes[i + 1] == 0xff, 0, 1); - - CHECK (2, 16, 11, bytes[i] == 0xff && bytes[i + 1] == bytes[1], 8, 1); - - CHECK (4, 32, 12, bytes[i] == 0xff && bytes[i + 1] == bytes[1] - && bytes[i + 2] == 0 && bytes[i + 3] == 0, 0, 0); - - CHECK (4, 32, 13, bytes[i] == 0 && bytes[i + 1] == bytes[1] - && bytes[i + 2] == 0xff && bytes[i + 3] == 0xff, 0, 1); - - CHECK (4, 32, 14, bytes[i] == 0xff && bytes[i + 1] == 0xff - && bytes[i + 2] == bytes[2] && bytes[i + 3] == 0, 0, 0); - - CHECK (4, 32, 15, bytes[i] == 0 && bytes[i + 1] == 0 - && bytes[i + 2] == bytes[2] && bytes[i + 3] == 0xff, 0, 1); - - CHECK (1, 8, 16, bytes[i] == bytes[0], 0, 0); - - CHECK (1, 64, 17, (bytes[i] == 0 || bytes[i] == 0xff) - && bytes[i] == bytes[(i + 8) % idx], 0, 0); - } - while (0); - - /* TODO: Currently the assembler cannot handle types 12 to 15. - And there is no way to specify cmode through the compiler. - Disable them till there is support in the assembler. */ - if (immtype == -1 - || (immtype >= 12 && immtype <= 15) - || immtype == 18) - return -1; - - - if (elementwidth) - *elementwidth = elsize; - - if (elementchar) - *elementchar = elchar; - - if (mvn) - *mvn = emvn; - - if (shift) - *shift = eshift; - - if (modconst) - { - unsigned HOST_WIDE_INT imm = 0; - - /* Un-invert bytes of recognized vector, if necessary. */ - if (invmask != 0) - for (i = 0; i < idx; i++) - bytes[i] ^= invmask; - - if (immtype == 17) - { - /* FIXME: Broken on 32-bit H_W_I hosts. */ - gcc_assert (sizeof (HOST_WIDE_INT) == 8); - - for (i = 0; i < 8; i++) - imm |= (unsigned HOST_WIDE_INT) (bytes[i] ? 0xff : 0) - << (i * BITS_PER_UNIT); - - *modconst = GEN_INT (imm); - } - else - { - unsigned HOST_WIDE_INT imm = 0; - - for (i = 0; i < elsize / BITS_PER_UNIT; i++) - imm |= (unsigned HOST_WIDE_INT) bytes[i] << (i * BITS_PER_UNIT); - - /* Construct 'abcdefgh' because the assembler cannot handle - generic constants. */ - gcc_assert (shift != NULL && mvn != NULL); - if (*mvn) - imm = ~imm; - imm = (imm >> *shift) & 0xff; - *modconst = GEN_INT (imm); - } - } - - return immtype; -#undef CHECK -} - -/* Return TRUE if rtx X is legal for use as either a AdvSIMD MOVI instruction - (or, implicitly, MVNI) immediate. Write back width per element - to *ELEMENTWIDTH, and a modified constant (whatever should be output - for a MOVI instruction) in *MODCONST. */ -int -aarch64_simd_immediate_valid_for_move (rtx op, enum machine_mode mode, - rtx *modconst, int *elementwidth, - unsigned char *elementchar, - int *mvn, int *shift) -{ - rtx tmpconst; - int tmpwidth; - unsigned char tmpwidthc; - int tmpmvn = 0, tmpshift = 0; - int retval = aarch64_simd_valid_immediate (op, mode, 0, &tmpconst, - &tmpwidth, &tmpwidthc, - &tmpmvn, &tmpshift); - - if (retval == -1) - return 0; - - if (modconst) - *modconst = tmpconst; - - if (elementwidth) - *elementwidth = tmpwidth; - - if (elementchar) - *elementchar = tmpwidthc; - - if (mvn) - *mvn = tmpmvn; - - if (shift) - *shift = tmpshift; - - return 1; -} - -static bool -aarch64_const_vec_all_same_int_p (rtx x, - HOST_WIDE_INT minval, - HOST_WIDE_INT maxval) -{ - HOST_WIDE_INT firstval; - int count, i; - - if (GET_CODE (x) != CONST_VECTOR - || GET_MODE_CLASS (GET_MODE (x)) != MODE_VECTOR_INT) - return false; - - firstval = INTVAL (CONST_VECTOR_ELT (x, 0)); - if (firstval < minval || firstval > maxval) - return false; - - count = CONST_VECTOR_NUNITS (x); - for (i = 1; i < count; i++) - if (INTVAL (CONST_VECTOR_ELT (x, i)) != firstval) - return false; - - return true; -} - -/* Check of immediate shift constants are within range. */ -bool -aarch64_simd_shift_imm_p (rtx x, enum machine_mode mode, bool left) -{ - int bit_width = GET_MODE_UNIT_SIZE (mode) * BITS_PER_UNIT; - if (left) - return aarch64_const_vec_all_same_int_p (x, 0, bit_width - 1); - else - return aarch64_const_vec_all_same_int_p (x, 1, bit_width); -} - -/* Return true if X is a uniform vector where all elements - are either the floating-point constant 0.0 or the - integer constant 0. */ -bool -aarch64_simd_imm_zero_p (rtx x, enum machine_mode mode) -{ - return x == CONST0_RTX (mode); -} - -bool -aarch64_simd_imm_scalar_p (rtx x, enum machine_mode mode ATTRIBUTE_UNUSED) -{ - HOST_WIDE_INT imm = INTVAL (x); - int i; - - for (i = 0; i < 8; i++) - { - unsigned int byte = imm & 0xff; - if (byte != 0xff && byte != 0) - return false; - imm >>= 8; - } - - return true; -} - -/* Return a const_int vector of VAL. */ -rtx -aarch64_simd_gen_const_vector_dup (enum machine_mode mode, int val) -{ - int nunits = GET_MODE_NUNITS (mode); - rtvec v = rtvec_alloc (nunits); - int i; - - for (i=0; i < nunits; i++) - RTVEC_ELT (v, i) = GEN_INT (val); - - return gen_rtx_CONST_VECTOR (mode, v); -} - -/* Construct and return a PARALLEL RTX vector. */ -rtx -aarch64_simd_vect_par_cnst_half (enum machine_mode mode, bool high) -{ - int nunits = GET_MODE_NUNITS (mode); - rtvec v = rtvec_alloc (nunits / 2); - int base = high ? nunits / 2 : 0; - rtx t1; - int i; - - for (i=0; i < nunits / 2; i++) - RTVEC_ELT (v, i) = GEN_INT (base + i); - - t1 = gen_rtx_PARALLEL (mode, v); - return t1; -} - -/* Bounds-check lanes. Ensure OPERAND lies between LOW (inclusive) and - HIGH (exclusive). */ -void -aarch64_simd_lane_bounds (rtx operand, HOST_WIDE_INT low, HOST_WIDE_INT high) -{ - HOST_WIDE_INT lane; - gcc_assert (GET_CODE (operand) == CONST_INT); - lane = INTVAL (operand); - - if (lane < low || lane >= high) - error ("lane out of range"); -} - -void -aarch64_simd_const_bounds (rtx operand, HOST_WIDE_INT low, HOST_WIDE_INT high) -{ - gcc_assert (GET_CODE (operand) == CONST_INT); - HOST_WIDE_INT lane = INTVAL (operand); - - if (lane < low || lane >= high) - error ("constant out of range"); -} - -/* Emit code to reinterpret one AdvSIMD type as another, - without altering bits. */ -void -aarch64_simd_reinterpret (rtx dest, rtx src) -{ - emit_move_insn (dest, gen_lowpart (GET_MODE (dest), src)); -} - -/* Emit code to place a AdvSIMD pair result in memory locations (with equal - registers). */ -void -aarch64_simd_emit_pair_result_insn (enum machine_mode mode, - rtx (*intfn) (rtx, rtx, rtx), rtx destaddr, - rtx op1) -{ - rtx mem = gen_rtx_MEM (mode, destaddr); - rtx tmp1 = gen_reg_rtx (mode); - rtx tmp2 = gen_reg_rtx (mode); - - emit_insn (intfn (tmp1, op1, tmp2)); - - emit_move_insn (mem, tmp1); - mem = adjust_address (mem, mode, GET_MODE_SIZE (mode)); - emit_move_insn (mem, tmp2); -} - -/* Return TRUE if OP is a valid vector addressing mode. */ -bool -aarch64_simd_mem_operand_p (rtx op) -{ - return MEM_P (op) && (GET_CODE (XEXP (op, 0)) == POST_INC - || GET_CODE (XEXP (op, 0)) == REG); -} - -/* Set up OPERANDS for a register copy from SRC to DEST, taking care - not to early-clobber SRC registers in the process. - - We assume that the operands described by SRC and DEST represent a - decomposed copy of OPERANDS[1] into OPERANDS[0]. COUNT is the - number of components into which the copy has been decomposed. */ -void -aarch64_simd_disambiguate_copy (rtx *operands, rtx *dest, - rtx *src, unsigned int count) -{ - unsigned int i; - - if (!reg_overlap_mentioned_p (operands[0], operands[1]) - || REGNO (operands[0]) < REGNO (operands[1])) - { - for (i = 0; i < count; i++) - { - operands[2 * i] = dest[i]; - operands[2 * i + 1] = src[i]; - } - } - else - { - for (i = 0; i < count; i++) - { - operands[2 * i] = dest[count - i - 1]; - operands[2 * i + 1] = src[count - i - 1]; - } - } -} - -/* Compute and return the length of aarch64_simd_mov<mode>, where <mode> is - one of VSTRUCT modes: OI, CI or XI. */ -int -aarch64_simd_attr_length_move (rtx insn) -{ - enum machine_mode mode; - - extract_insn_cached (insn); - - if (REG_P (recog_data.operand[0]) && REG_P (recog_data.operand[1])) - { - mode = GET_MODE (recog_data.operand[0]); - switch (mode) - { - case OImode: - return 8; - case CImode: - return 12; - case XImode: - return 16; - default: - gcc_unreachable (); - } - } - return 4; -} - -/* Implement target hook TARGET_VECTOR_ALIGNMENT. The AAPCS64 sets the maximum - alignment of a vector to 128 bits. */ -static HOST_WIDE_INT -aarch64_simd_vector_alignment (const_tree type) -{ - HOST_WIDE_INT align = tree_low_cst (TYPE_SIZE (type), 0); - return MIN (align, 128); -} - -/* Implement target hook TARGET_VECTORIZE_VECTOR_ALIGNMENT_REACHABLE. */ -static bool -aarch64_simd_vector_alignment_reachable (const_tree type, bool is_packed) -{ - if (is_packed) - return false; - - /* We guarantee alignment for vectors up to 128-bits. */ - if (tree_int_cst_compare (TYPE_SIZE (type), - bitsize_int (BIGGEST_ALIGNMENT)) > 0) - return false; - - /* Vectors whose size is <= BIGGEST_ALIGNMENT are naturally aligned. */ - return true; -} - -/* If VALS is a vector constant that can be loaded into a register - using DUP, generate instructions to do so and return an RTX to - assign to the register. Otherwise return NULL_RTX. */ -static rtx -aarch64_simd_dup_constant (rtx vals) -{ - enum machine_mode mode = GET_MODE (vals); - enum machine_mode inner_mode = GET_MODE_INNER (mode); - int n_elts = GET_MODE_NUNITS (mode); - bool all_same = true; - rtx x; - int i; - - if (GET_CODE (vals) != CONST_VECTOR) - return NULL_RTX; - - for (i = 1; i < n_elts; ++i) - { - x = CONST_VECTOR_ELT (vals, i); - if (!rtx_equal_p (x, CONST_VECTOR_ELT (vals, 0))) - all_same = false; - } - - if (!all_same) - return NULL_RTX; - - /* We can load this constant by using DUP and a constant in a - single ARM register. This will be cheaper than a vector - load. */ - x = copy_to_mode_reg (inner_mode, CONST_VECTOR_ELT (vals, 0)); - return gen_rtx_VEC_DUPLICATE (mode, x); -} - - -/* Generate code to load VALS, which is a PARALLEL containing only - constants (for vec_init) or CONST_VECTOR, efficiently into a - register. Returns an RTX to copy into the register, or NULL_RTX - for a PARALLEL that can not be converted into a CONST_VECTOR. */ -static rtx -aarch64_simd_make_constant (rtx vals) -{ - enum machine_mode mode = GET_MODE (vals); - rtx const_dup; - rtx const_vec = NULL_RTX; - int n_elts = GET_MODE_NUNITS (mode); - int n_const = 0; - int i; - - if (GET_CODE (vals) == CONST_VECTOR) - const_vec = vals; - else if (GET_CODE (vals) == PARALLEL) - { - /* A CONST_VECTOR must contain only CONST_INTs and - CONST_DOUBLEs, but CONSTANT_P allows more (e.g. SYMBOL_REF). - Only store valid constants in a CONST_VECTOR. */ - for (i = 0; i < n_elts; ++i) - { - rtx x = XVECEXP (vals, 0, i); - if (CONST_INT_P (x) || CONST_DOUBLE_P (x)) - n_const++; - } - if (n_const == n_elts) - const_vec = gen_rtx_CONST_VECTOR (mode, XVEC (vals, 0)); - } - else - gcc_unreachable (); - - if (const_vec != NULL_RTX - && aarch64_simd_immediate_valid_for_move (const_vec, mode, NULL, NULL, - NULL, NULL, NULL)) - /* Load using MOVI/MVNI. */ - return const_vec; - else if ((const_dup = aarch64_simd_dup_constant (vals)) != NULL_RTX) - /* Loaded using DUP. */ - return const_dup; - else if (const_vec != NULL_RTX) - /* Load from constant pool. We can not take advantage of single-cycle - LD1 because we need a PC-relative addressing mode. */ - return const_vec; - else - /* A PARALLEL containing something not valid inside CONST_VECTOR. - We can not construct an initializer. */ - return NULL_RTX; -} - -void -aarch64_expand_vector_init (rtx target, rtx vals) -{ - enum machine_mode mode = GET_MODE (target); - enum machine_mode inner_mode = GET_MODE_INNER (mode); - int n_elts = GET_MODE_NUNITS (mode); - int n_var = 0, one_var = -1; - bool all_same = true; - rtx x, mem; - int i; - - x = XVECEXP (vals, 0, 0); - if (!CONST_INT_P (x) && !CONST_DOUBLE_P (x)) - n_var = 1, one_var = 0; - - for (i = 1; i < n_elts; ++i) - { - x = XVECEXP (vals, 0, i); - if (!CONST_INT_P (x) && !CONST_DOUBLE_P (x)) - ++n_var, one_var = i; - - if (!rtx_equal_p (x, XVECEXP (vals, 0, 0))) - all_same = false; - } - - if (n_var == 0) - { - rtx constant = aarch64_simd_make_constant (vals); - if (constant != NULL_RTX) - { - emit_move_insn (target, constant); - return; - } - } - - /* Splat a single non-constant element if we can. */ - if (all_same) - { - x = copy_to_mode_reg (inner_mode, XVECEXP (vals, 0, 0)); - aarch64_emit_move (target, gen_rtx_VEC_DUPLICATE (mode, x)); - return; - } - - /* One field is non-constant. Load constant then overwrite varying - field. This is more efficient than using the stack. */ - if (n_var == 1) - { - rtx copy = copy_rtx (vals); - rtx index = GEN_INT (one_var); - enum insn_code icode; - - /* Load constant part of vector, substitute neighboring value for - varying element. */ - XVECEXP (copy, 0, one_var) = XVECEXP (vals, 0, one_var ^ 1); - aarch64_expand_vector_init (target, copy); - - /* Insert variable. */ - x = copy_to_mode_reg (inner_mode, XVECEXP (vals, 0, one_var)); - icode = optab_handler (vec_set_optab, mode); - gcc_assert (icode != CODE_FOR_nothing); - emit_insn (GEN_FCN (icode) (target, x, index)); - return; - } - - /* Construct the vector in memory one field at a time - and load the whole vector. */ - mem = assign_stack_temp (mode, GET_MODE_SIZE (mode)); - for (i = 0; i < n_elts; i++) - emit_move_insn (adjust_address_nv (mem, inner_mode, - i * GET_MODE_SIZE (inner_mode)), - XVECEXP (vals, 0, i)); - emit_move_insn (target, mem); - -} - -static unsigned HOST_WIDE_INT -aarch64_shift_truncation_mask (enum machine_mode mode) -{ - return - (aarch64_vector_mode_supported_p (mode) - || aarch64_vect_struct_mode_p (mode)) ? 0 : (GET_MODE_BITSIZE (mode) - 1); -} - -#ifndef TLS_SECTION_ASM_FLAG -#define TLS_SECTION_ASM_FLAG 'T' -#endif - -void -aarch64_elf_asm_named_section (const char *name, unsigned int flags, - tree decl ATTRIBUTE_UNUSED) -{ - char flagchars[10], *f = flagchars; - - /* If we have already declared this section, we can use an - abbreviated form to switch back to it -- unless this section is - part of a COMDAT groups, in which case GAS requires the full - declaration every time. */ - if (!(HAVE_COMDAT_GROUP && (flags & SECTION_LINKONCE)) - && (flags & SECTION_DECLARED)) - { - fprintf (asm_out_file, "\t.section\t%s\n", name); - return; - } - - if (!(flags & SECTION_DEBUG)) - *f++ = 'a'; - if (flags & SECTION_WRITE) - *f++ = 'w'; - if (flags & SECTION_CODE) - *f++ = 'x'; - if (flags & SECTION_SMALL) - *f++ = 's'; - if (flags & SECTION_MERGE) - *f++ = 'M'; - if (flags & SECTION_STRINGS) - *f++ = 'S'; - if (flags & SECTION_TLS) - *f++ = TLS_SECTION_ASM_FLAG; - if (HAVE_COMDAT_GROUP && (flags & SECTION_LINKONCE)) - *f++ = 'G'; - *f = '\0'; - - fprintf (asm_out_file, "\t.section\t%s,\"%s\"", name, flagchars); - - if (!(flags & SECTION_NOTYPE)) - { - const char *type; - const char *format; - - if (flags & SECTION_BSS) - type = "nobits"; - else - type = "progbits"; - -#ifdef TYPE_OPERAND_FMT - format = "," TYPE_OPERAND_FMT; -#else - format = ",@%s"; -#endif - - fprintf (asm_out_file, format, type); - - if (flags & SECTION_ENTSIZE) - fprintf (asm_out_file, ",%d", flags & SECTION_ENTSIZE); - if (HAVE_COMDAT_GROUP && (flags & SECTION_LINKONCE)) - { - if (TREE_CODE (decl) == IDENTIFIER_NODE) - fprintf (asm_out_file, ",%s,comdat", IDENTIFIER_POINTER (decl)); - else - fprintf (asm_out_file, ",%s,comdat", - IDENTIFIER_POINTER (DECL_COMDAT_GROUP (decl))); - } - } - - putc ('\n', asm_out_file); -} - -/* Select a format to encode pointers in exception handling data. */ -int -aarch64_asm_preferred_eh_data_format (int code ATTRIBUTE_UNUSED, int global) -{ - int type; - switch (aarch64_cmodel) - { - case AARCH64_CMODEL_TINY: - case AARCH64_CMODEL_TINY_PIC: - case AARCH64_CMODEL_SMALL: - case AARCH64_CMODEL_SMALL_PIC: - /* text+got+data < 4Gb. 4-byte signed relocs are sufficient - for everything. */ - type = DW_EH_PE_sdata4; - break; - default: - /* No assumptions here. 8-byte relocs required. */ - type = DW_EH_PE_sdata8; - break; - } - return (global ? DW_EH_PE_indirect : 0) | DW_EH_PE_pcrel | type; -} - -/* Emit load exclusive. */ - -static void -aarch64_emit_load_exclusive (enum machine_mode mode, rtx rval, - rtx mem, rtx model_rtx) -{ - rtx (*gen) (rtx, rtx, rtx); - - switch (mode) - { - case QImode: gen = gen_aarch64_load_exclusiveqi; break; - case HImode: gen = gen_aarch64_load_exclusivehi; break; - case SImode: gen = gen_aarch64_load_exclusivesi; break; - case DImode: gen = gen_aarch64_load_exclusivedi; break; - default: - gcc_unreachable (); - } - - emit_insn (gen (rval, mem, model_rtx)); -} - -/* Emit store exclusive. */ - -static void -aarch64_emit_store_exclusive (enum machine_mode mode, rtx bval, - rtx rval, rtx mem, rtx model_rtx) -{ - rtx (*gen) (rtx, rtx, rtx, rtx); - - switch (mode) - { - case QImode: gen = gen_aarch64_store_exclusiveqi; break; - case HImode: gen = gen_aarch64_store_exclusivehi; break; - case SImode: gen = gen_aarch64_store_exclusivesi; break; - case DImode: gen = gen_aarch64_store_exclusivedi; break; - default: - gcc_unreachable (); - } - - emit_insn (gen (bval, rval, mem, model_rtx)); -} - -/* Mark the previous jump instruction as unlikely. */ - -static void -aarch64_emit_unlikely_jump (rtx insn) -{ - rtx very_unlikely = GEN_INT (REG_BR_PROB_BASE / 100 - 1); - - insn = emit_jump_insn (insn); - add_reg_note (insn, REG_BR_PROB, very_unlikely); -} - -/* Expand a compare and swap pattern. */ - -void -aarch64_expand_compare_and_swap (rtx operands[]) -{ - rtx bval, rval, mem, oldval, newval, is_weak, mod_s, mod_f, x; - enum machine_mode mode, cmp_mode; - rtx (*gen) (rtx, rtx, rtx, rtx, rtx, rtx, rtx); - - bval = operands[0]; - rval = operands[1]; - mem = operands[2]; - oldval = operands[3]; - newval = operands[4]; - is_weak = operands[5]; - mod_s = operands[6]; - mod_f = operands[7]; - mode = GET_MODE (mem); - cmp_mode = mode; - - /* Normally the succ memory model must be stronger than fail, but in the - unlikely event of fail being ACQUIRE and succ being RELEASE we need to - promote succ to ACQ_REL so that we don't lose the acquire semantics. */ - - if (INTVAL (mod_f) == MEMMODEL_ACQUIRE - && INTVAL (mod_s) == MEMMODEL_RELEASE) - mod_s = GEN_INT (MEMMODEL_ACQ_REL); - - switch (mode) - { - case QImode: - case HImode: - /* For short modes, we're going to perform the comparison in SImode, - so do the zero-extension now. */ - cmp_mode = SImode; - rval = gen_reg_rtx (SImode); - oldval = convert_modes (SImode, mode, oldval, true); - /* Fall through. */ - - case SImode: - case DImode: - /* Force the value into a register if needed. */ - if (!aarch64_plus_operand (oldval, mode)) - oldval = force_reg (cmp_mode, oldval); - break; - - default: - gcc_unreachable (); - } - - switch (mode) - { - case QImode: gen = gen_atomic_compare_and_swapqi_1; break; - case HImode: gen = gen_atomic_compare_and_swaphi_1; break; - case SImode: gen = gen_atomic_compare_and_swapsi_1; break; - case DImode: gen = gen_atomic_compare_and_swapdi_1; break; - default: - gcc_unreachable (); - } - - emit_insn (gen (rval, mem, oldval, newval, is_weak, mod_s, mod_f)); - - if (mode == QImode || mode == HImode) - emit_move_insn (operands[1], gen_lowpart (mode, rval)); - - x = gen_rtx_REG (CCmode, CC_REGNUM); - x = gen_rtx_EQ (SImode, x, const0_rtx); - emit_insn (gen_rtx_SET (VOIDmode, bval, x)); -} - -/* Split a compare and swap pattern. */ - -void -aarch64_split_compare_and_swap (rtx operands[]) -{ - rtx rval, mem, oldval, newval, scratch; - enum machine_mode mode; - bool is_weak; - rtx label1, label2, x, cond; - - rval = operands[0]; - mem = operands[1]; - oldval = operands[2]; - newval = operands[3]; - is_weak = (operands[4] != const0_rtx); - scratch = operands[7]; - mode = GET_MODE (mem); - - label1 = NULL_RTX; - if (!is_weak) - { - label1 = gen_label_rtx (); - emit_label (label1); - } - label2 = gen_label_rtx (); - - aarch64_emit_load_exclusive (mode, rval, mem, operands[5]); - - cond = aarch64_gen_compare_reg (NE, rval, oldval); - x = gen_rtx_NE (VOIDmode, cond, const0_rtx); - x = gen_rtx_IF_THEN_ELSE (VOIDmode, x, - gen_rtx_LABEL_REF (Pmode, label2), pc_rtx); - aarch64_emit_unlikely_jump (gen_rtx_SET (VOIDmode, pc_rtx, x)); - - aarch64_emit_store_exclusive (mode, scratch, mem, newval, operands[5]); - - if (!is_weak) - { - x = gen_rtx_NE (VOIDmode, scratch, const0_rtx); - x = gen_rtx_IF_THEN_ELSE (VOIDmode, x, - gen_rtx_LABEL_REF (Pmode, label1), pc_rtx); - aarch64_emit_unlikely_jump (gen_rtx_SET (VOIDmode, pc_rtx, x)); - } - else - { - cond = gen_rtx_REG (CCmode, CC_REGNUM); - x = gen_rtx_COMPARE (CCmode, scratch, const0_rtx); - emit_insn (gen_rtx_SET (VOIDmode, cond, x)); - } - - emit_label (label2); -} - -/* Split an atomic operation. */ - -void -aarch64_split_atomic_op (enum rtx_code code, rtx old_out, rtx new_out, rtx mem, - rtx value, rtx model_rtx, rtx cond) -{ - enum machine_mode mode = GET_MODE (mem); - enum machine_mode wmode = (mode == DImode ? DImode : SImode); - rtx label, x; - - label = gen_label_rtx (); - emit_label (label); - - if (new_out) - new_out = gen_lowpart (wmode, new_out); - if (old_out) - old_out = gen_lowpart (wmode, old_out); - else - old_out = new_out; - value = simplify_gen_subreg (wmode, value, mode, 0); - - aarch64_emit_load_exclusive (mode, old_out, mem, model_rtx); - - switch (code) - { - case SET: - new_out = value; - break; - - case NOT: - x = gen_rtx_AND (wmode, old_out, value); - emit_insn (gen_rtx_SET (VOIDmode, new_out, x)); - x = gen_rtx_NOT (wmode, new_out); - emit_insn (gen_rtx_SET (VOIDmode, new_out, x)); - break; - - case MINUS: - if (CONST_INT_P (value)) - { - value = GEN_INT (-INTVAL (value)); - code = PLUS; - } - /* Fall through. */ - - default: - x = gen_rtx_fmt_ee (code, wmode, old_out, value); - emit_insn (gen_rtx_SET (VOIDmode, new_out, x)); - break; - } - - aarch64_emit_store_exclusive (mode, cond, mem, - gen_lowpart (mode, new_out), model_rtx); - - x = gen_rtx_NE (VOIDmode, cond, const0_rtx); - x = gen_rtx_IF_THEN_ELSE (VOIDmode, x, - gen_rtx_LABEL_REF (Pmode, label), pc_rtx); - aarch64_emit_unlikely_jump (gen_rtx_SET (VOIDmode, pc_rtx, x)); -} - -static void -aarch64_print_extension (void) -{ - const struct aarch64_option_extension *opt = NULL; - - for (opt = all_extensions; opt->name != NULL; opt++) - if ((aarch64_isa_flags & opt->flags_on) == opt->flags_on) - asm_fprintf (asm_out_file, "+%s", opt->name); - - asm_fprintf (asm_out_file, "\n"); -} - -static void -aarch64_start_file (void) -{ - if (selected_arch) - { - asm_fprintf (asm_out_file, "\t.arch %s", selected_arch->name); - aarch64_print_extension (); - } - else if (selected_cpu) - { - asm_fprintf (asm_out_file, "\t.cpu %s", selected_cpu->name); - aarch64_print_extension (); - } - default_file_start(); -} - -/* Target hook for c_mode_for_suffix. */ -static enum machine_mode -aarch64_c_mode_for_suffix (char suffix) -{ - if (suffix == 'q') - return TFmode; - - return VOIDmode; -} - -/* We can only represent floating point constants which will fit in - "quarter-precision" values. These values are characterised by - a sign bit, a 4-bit mantissa and a 3-bit exponent. And are given - by: - - (-1)^s * (n/16) * 2^r - - Where: - 's' is the sign bit. - 'n' is an integer in the range 16 <= n <= 31. - 'r' is an integer in the range -3 <= r <= 4. */ - -/* Return true iff X can be represented by a quarter-precision - floating point immediate operand X. Note, we cannot represent 0.0. */ -bool -aarch64_float_const_representable_p (rtx x) -{ - /* This represents our current view of how many bits - make up the mantissa. */ - int point_pos = 2 * HOST_BITS_PER_WIDE_INT - 1; - int exponent; - unsigned HOST_WIDE_INT mantissa, mask; - HOST_WIDE_INT m1, m2; - REAL_VALUE_TYPE r, m; - - if (!CONST_DOUBLE_P (x)) - return false; - - REAL_VALUE_FROM_CONST_DOUBLE (r, x); - - /* We cannot represent infinities, NaNs or +/-zero. We won't - know if we have +zero until we analyse the mantissa, but we - can reject the other invalid values. */ - if (REAL_VALUE_ISINF (r) || REAL_VALUE_ISNAN (r) - || REAL_VALUE_MINUS_ZERO (r)) - return false; - - /* Extract exponent. */ - r = real_value_abs (&r); - exponent = REAL_EXP (&r); - - /* For the mantissa, we expand into two HOST_WIDE_INTS, apart from the - highest (sign) bit, with a fixed binary point at bit point_pos. - m1 holds the low part of the mantissa, m2 the high part. - WARNING: If we ever have a representation using more than 2 * H_W_I - 1 - bits for the mantissa, this can fail (low bits will be lost). */ - real_ldexp (&m, &r, point_pos - exponent); - REAL_VALUE_TO_INT (&m1, &m2, m); - - /* If the low part of the mantissa has bits set we cannot represent - the value. */ - if (m1 != 0) - return false; - /* We have rejected the lower HOST_WIDE_INT, so update our - understanding of how many bits lie in the mantissa and - look only at the high HOST_WIDE_INT. */ - mantissa = m2; - point_pos -= HOST_BITS_PER_WIDE_INT; - - /* We can only represent values with a mantissa of the form 1.xxxx. */ - mask = ((unsigned HOST_WIDE_INT)1 << (point_pos - 5)) - 1; - if ((mantissa & mask) != 0) - return false; - - /* Having filtered unrepresentable values, we may now remove all - but the highest 5 bits. */ - mantissa >>= point_pos - 5; - - /* We cannot represent the value 0.0, so reject it. This is handled - elsewhere. */ - if (mantissa == 0) - return false; - - /* Then, as bit 4 is always set, we can mask it off, leaving - the mantissa in the range [0, 15]. */ - mantissa &= ~(1 << 4); - gcc_assert (mantissa <= 15); - - /* GCC internally does not use IEEE754-like encoding (where normalized - significands are in the range [1, 2). GCC uses [0.5, 1) (see real.c). - Our mantissa values are shifted 4 places to the left relative to - normalized IEEE754 so we must modify the exponent returned by REAL_EXP - by 5 places to correct for GCC's representation. */ - exponent = 5 - exponent; - - return (exponent >= 0 && exponent <= 7); -} - -char* -aarch64_output_simd_mov_immediate (rtx *const_vector, - enum machine_mode mode, - unsigned width) -{ - int is_valid; - unsigned char widthc; - int lane_width_bits; - static char templ[40]; - int shift = 0, mvn = 0; - const char *mnemonic; - unsigned int lane_count = 0; - - is_valid = - aarch64_simd_immediate_valid_for_move (*const_vector, mode, - const_vector, &lane_width_bits, - &widthc, &mvn, &shift); - gcc_assert (is_valid); - - mode = GET_MODE_INNER (mode); - if (mode == SFmode || mode == DFmode) - { - bool zero_p = - aarch64_float_const_zero_rtx_p (*const_vector); - gcc_assert (shift == 0); - mnemonic = zero_p ? "movi" : "fmov"; - } - else - mnemonic = mvn ? "mvni" : "movi"; - - gcc_assert (lane_width_bits != 0); - lane_count = width / lane_width_bits; - - if (lane_count == 1) - snprintf (templ, sizeof (templ), "%s\t%%d0, %%1", mnemonic); - else if (shift) - snprintf (templ, sizeof (templ), "%s\t%%0.%d%c, %%1, lsl %d", - mnemonic, lane_count, widthc, shift); - else - snprintf (templ, sizeof (templ), "%s\t%%0.%d%c, %%1", - mnemonic, lane_count, widthc); - return templ; -} - -/* Split operands into moves from op[1] + op[2] into op[0]. */ - -void -aarch64_split_combinev16qi (rtx operands[3]) -{ - unsigned int dest = REGNO (operands[0]); - unsigned int src1 = REGNO (operands[1]); - unsigned int src2 = REGNO (operands[2]); - enum machine_mode halfmode = GET_MODE (operands[1]); - unsigned int halfregs = HARD_REGNO_NREGS (src1, halfmode); - rtx destlo, desthi; - - gcc_assert (halfmode == V16QImode); - - if (src1 == dest && src2 == dest + halfregs) - { - /* No-op move. Can't split to nothing; emit something. */ - emit_note (NOTE_INSN_DELETED); - return; - } - - /* Preserve register attributes for variable tracking. */ - destlo = gen_rtx_REG_offset (operands[0], halfmode, dest, 0); - desthi = gen_rtx_REG_offset (operands[0], halfmode, dest + halfregs, - GET_MODE_SIZE (halfmode)); - - /* Special case of reversed high/low parts. */ - if (reg_overlap_mentioned_p (operands[2], destlo) - && reg_overlap_mentioned_p (operands[1], desthi)) - { - emit_insn (gen_xorv16qi3 (operands[1], operands[1], operands[2])); - emit_insn (gen_xorv16qi3 (operands[2], operands[1], operands[2])); - emit_insn (gen_xorv16qi3 (operands[1], operands[1], operands[2])); - } - else if (!reg_overlap_mentioned_p (operands[2], destlo)) - { - /* Try to avoid unnecessary moves if part of the result - is in the right place already. */ - if (src1 != dest) - emit_move_insn (destlo, operands[1]); - if (src2 != dest + halfregs) - emit_move_insn (desthi, operands[2]); - } - else - { - if (src2 != dest + halfregs) - emit_move_insn (desthi, operands[2]); - if (src1 != dest) - emit_move_insn (destlo, operands[1]); - } -} - -/* vec_perm support. */ - -#define MAX_VECT_LEN 16 - -struct expand_vec_perm_d -{ - rtx target, op0, op1; - unsigned char perm[MAX_VECT_LEN]; - enum machine_mode vmode; - unsigned char nelt; - bool one_vector_p; - bool testing_p; -}; - -/* Generate a variable permutation. */ - -static void -aarch64_expand_vec_perm_1 (rtx target, rtx op0, rtx op1, rtx sel) -{ - enum machine_mode vmode = GET_MODE (target); - bool one_vector_p = rtx_equal_p (op0, op1); - - gcc_checking_assert (vmode == V8QImode || vmode == V16QImode); - gcc_checking_assert (GET_MODE (op0) == vmode); - gcc_checking_assert (GET_MODE (op1) == vmode); - gcc_checking_assert (GET_MODE (sel) == vmode); - gcc_checking_assert (TARGET_SIMD); - - if (one_vector_p) - { - if (vmode == V8QImode) - { - /* Expand the argument to a V16QI mode by duplicating it. */ - rtx pair = gen_reg_rtx (V16QImode); - emit_insn (gen_aarch64_combinev8qi (pair, op0, op0)); - emit_insn (gen_aarch64_tbl1v8qi (target, pair, sel)); - } - else - { - emit_insn (gen_aarch64_tbl1v16qi (target, op0, sel)); - } - } - else - { - rtx pair; - - if (vmode == V8QImode) - { - pair = gen_reg_rtx (V16QImode); - emit_insn (gen_aarch64_combinev8qi (pair, op0, op1)); - emit_insn (gen_aarch64_tbl1v8qi (target, pair, sel)); - } - else - { - pair = gen_reg_rtx (OImode); - emit_insn (gen_aarch64_combinev16qi (pair, op0, op1)); - emit_insn (gen_aarch64_tbl2v16qi (target, pair, sel)); - } - } -} - -void -aarch64_expand_vec_perm (rtx target, rtx op0, rtx op1, rtx sel) -{ - enum machine_mode vmode = GET_MODE (target); - unsigned int i, nelt = GET_MODE_NUNITS (vmode); - bool one_vector_p = rtx_equal_p (op0, op1); - rtx rmask[MAX_VECT_LEN], mask; - - gcc_checking_assert (!BYTES_BIG_ENDIAN); - - /* The TBL instruction does not use a modulo index, so we must take care - of that ourselves. */ - mask = GEN_INT (one_vector_p ? nelt - 1 : 2 * nelt - 1); - for (i = 0; i < nelt; ++i) - rmask[i] = mask; - mask = gen_rtx_CONST_VECTOR (vmode, gen_rtvec_v (nelt, rmask)); - sel = expand_simple_binop (vmode, AND, sel, mask, NULL, 0, OPTAB_LIB_WIDEN); - - aarch64_expand_vec_perm_1 (target, op0, op1, sel); -} - -/* Recognize patterns suitable for the TRN instructions. */ -static bool -aarch64_evpc_trn (struct expand_vec_perm_d *d) -{ - unsigned int i, odd, mask, nelt = d->nelt; - rtx out, in0, in1, x; - rtx (*gen) (rtx, rtx, rtx); - enum machine_mode vmode = d->vmode; - - if (GET_MODE_UNIT_SIZE (vmode) > 8) - return false; - - /* Note that these are little-endian tests. - We correct for big-endian later. */ - if (d->perm[0] == 0) - odd = 0; - else if (d->perm[0] == 1) - odd = 1; - else - return false; - mask = (d->one_vector_p ? nelt - 1 : 2 * nelt - 1); - - for (i = 0; i < nelt; i += 2) - { - if (d->perm[i] != i + odd) - return false; - if (d->perm[i + 1] != ((i + nelt + odd) & mask)) - return false; - } - - /* Success! */ - if (d->testing_p) - return true; - - in0 = d->op0; - in1 = d->op1; - if (BYTES_BIG_ENDIAN) - { - x = in0, in0 = in1, in1 = x; - odd = !odd; - } - out = d->target; - - if (odd) - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_trn2v16qi; break; - case V8QImode: gen = gen_aarch64_trn2v8qi; break; - case V8HImode: gen = gen_aarch64_trn2v8hi; break; - case V4HImode: gen = gen_aarch64_trn2v4hi; break; - case V4SImode: gen = gen_aarch64_trn2v4si; break; - case V2SImode: gen = gen_aarch64_trn2v2si; break; - case V2DImode: gen = gen_aarch64_trn2v2di; break; - case V4SFmode: gen = gen_aarch64_trn2v4sf; break; - case V2SFmode: gen = gen_aarch64_trn2v2sf; break; - case V2DFmode: gen = gen_aarch64_trn2v2df; break; - default: - return false; - } - } - else - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_trn1v16qi; break; - case V8QImode: gen = gen_aarch64_trn1v8qi; break; - case V8HImode: gen = gen_aarch64_trn1v8hi; break; - case V4HImode: gen = gen_aarch64_trn1v4hi; break; - case V4SImode: gen = gen_aarch64_trn1v4si; break; - case V2SImode: gen = gen_aarch64_trn1v2si; break; - case V2DImode: gen = gen_aarch64_trn1v2di; break; - case V4SFmode: gen = gen_aarch64_trn1v4sf; break; - case V2SFmode: gen = gen_aarch64_trn1v2sf; break; - case V2DFmode: gen = gen_aarch64_trn1v2df; break; - default: - return false; - } - } - - emit_insn (gen (out, in0, in1)); - return true; -} - -/* Recognize patterns suitable for the UZP instructions. */ -static bool -aarch64_evpc_uzp (struct expand_vec_perm_d *d) -{ - unsigned int i, odd, mask, nelt = d->nelt; - rtx out, in0, in1, x; - rtx (*gen) (rtx, rtx, rtx); - enum machine_mode vmode = d->vmode; - - if (GET_MODE_UNIT_SIZE (vmode) > 8) - return false; - - /* Note that these are little-endian tests. - We correct for big-endian later. */ - if (d->perm[0] == 0) - odd = 0; - else if (d->perm[0] == 1) - odd = 1; - else - return false; - mask = (d->one_vector_p ? nelt - 1 : 2 * nelt - 1); - - for (i = 0; i < nelt; i++) - { - unsigned elt = (i * 2 + odd) & mask; - if (d->perm[i] != elt) - return false; - } - - /* Success! */ - if (d->testing_p) - return true; - - in0 = d->op0; - in1 = d->op1; - if (BYTES_BIG_ENDIAN) - { - x = in0, in0 = in1, in1 = x; - odd = !odd; - } - out = d->target; - - if (odd) - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_uzp2v16qi; break; - case V8QImode: gen = gen_aarch64_uzp2v8qi; break; - case V8HImode: gen = gen_aarch64_uzp2v8hi; break; - case V4HImode: gen = gen_aarch64_uzp2v4hi; break; - case V4SImode: gen = gen_aarch64_uzp2v4si; break; - case V2SImode: gen = gen_aarch64_uzp2v2si; break; - case V2DImode: gen = gen_aarch64_uzp2v2di; break; - case V4SFmode: gen = gen_aarch64_uzp2v4sf; break; - case V2SFmode: gen = gen_aarch64_uzp2v2sf; break; - case V2DFmode: gen = gen_aarch64_uzp2v2df; break; - default: - return false; - } - } - else - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_uzp1v16qi; break; - case V8QImode: gen = gen_aarch64_uzp1v8qi; break; - case V8HImode: gen = gen_aarch64_uzp1v8hi; break; - case V4HImode: gen = gen_aarch64_uzp1v4hi; break; - case V4SImode: gen = gen_aarch64_uzp1v4si; break; - case V2SImode: gen = gen_aarch64_uzp1v2si; break; - case V2DImode: gen = gen_aarch64_uzp1v2di; break; - case V4SFmode: gen = gen_aarch64_uzp1v4sf; break; - case V2SFmode: gen = gen_aarch64_uzp1v2sf; break; - case V2DFmode: gen = gen_aarch64_uzp1v2df; break; - default: - return false; - } - } - - emit_insn (gen (out, in0, in1)); - return true; -} - -/* Recognize patterns suitable for the ZIP instructions. */ -static bool -aarch64_evpc_zip (struct expand_vec_perm_d *d) -{ - unsigned int i, high, mask, nelt = d->nelt; - rtx out, in0, in1, x; - rtx (*gen) (rtx, rtx, rtx); - enum machine_mode vmode = d->vmode; - - if (GET_MODE_UNIT_SIZE (vmode) > 8) - return false; - - /* Note that these are little-endian tests. - We correct for big-endian later. */ - high = nelt / 2; - if (d->perm[0] == high) - /* Do Nothing. */ - ; - else if (d->perm[0] == 0) - high = 0; - else - return false; - mask = (d->one_vector_p ? nelt - 1 : 2 * nelt - 1); - - for (i = 0; i < nelt / 2; i++) - { - unsigned elt = (i + high) & mask; - if (d->perm[i * 2] != elt) - return false; - elt = (elt + nelt) & mask; - if (d->perm[i * 2 + 1] != elt) - return false; - } - - /* Success! */ - if (d->testing_p) - return true; - - in0 = d->op0; - in1 = d->op1; - if (BYTES_BIG_ENDIAN) - { - x = in0, in0 = in1, in1 = x; - high = !high; - } - out = d->target; - - if (high) - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_zip2v16qi; break; - case V8QImode: gen = gen_aarch64_zip2v8qi; break; - case V8HImode: gen = gen_aarch64_zip2v8hi; break; - case V4HImode: gen = gen_aarch64_zip2v4hi; break; - case V4SImode: gen = gen_aarch64_zip2v4si; break; - case V2SImode: gen = gen_aarch64_zip2v2si; break; - case V2DImode: gen = gen_aarch64_zip2v2di; break; - case V4SFmode: gen = gen_aarch64_zip2v4sf; break; - case V2SFmode: gen = gen_aarch64_zip2v2sf; break; - case V2DFmode: gen = gen_aarch64_zip2v2df; break; - default: - return false; - } - } - else - { - switch (vmode) - { - case V16QImode: gen = gen_aarch64_zip1v16qi; break; - case V8QImode: gen = gen_aarch64_zip1v8qi; break; - case V8HImode: gen = gen_aarch64_zip1v8hi; break; - case V4HImode: gen = gen_aarch64_zip1v4hi; break; - case V4SImode: gen = gen_aarch64_zip1v4si; break; - case V2SImode: gen = gen_aarch64_zip1v2si; break; - case V2DImode: gen = gen_aarch64_zip1v2di; break; - case V4SFmode: gen = gen_aarch64_zip1v4sf; break; - case V2SFmode: gen = gen_aarch64_zip1v2sf; break; - case V2DFmode: gen = gen_aarch64_zip1v2df; break; - default: - return false; - } - } - - emit_insn (gen (out, in0, in1)); - return true; -} - -static bool -aarch64_evpc_tbl (struct expand_vec_perm_d *d) -{ - rtx rperm[MAX_VECT_LEN], sel; - enum machine_mode vmode = d->vmode; - unsigned int i, nelt = d->nelt; - - /* TODO: ARM's TBL indexing is little-endian. In order to handle GCC's - numbering of elements for big-endian, we must reverse the order. */ - if (BYTES_BIG_ENDIAN) - return false; - - if (d->testing_p) - return true; - - /* Generic code will try constant permutation twice. Once with the - original mode and again with the elements lowered to QImode. - So wait and don't do the selector expansion ourselves. */ - if (vmode != V8QImode && vmode != V16QImode) - return false; - - for (i = 0; i < nelt; ++i) - rperm[i] = GEN_INT (d->perm[i]); - sel = gen_rtx_CONST_VECTOR (vmode, gen_rtvec_v (nelt, rperm)); - sel = force_reg (vmode, sel); - - aarch64_expand_vec_perm_1 (d->target, d->op0, d->op1, sel); - return true; -} - -static bool -aarch64_expand_vec_perm_const_1 (struct expand_vec_perm_d *d) -{ - /* The pattern matching functions above are written to look for a small - number to begin the sequence (0, 1, N/2). If we begin with an index - from the second operand, we can swap the operands. */ - if (d->perm[0] >= d->nelt) - { - unsigned i, nelt = d->nelt; - rtx x; - - for (i = 0; i < nelt; ++i) - d->perm[i] = (d->perm[i] + nelt) & (2 * nelt - 1); - - x = d->op0; - d->op0 = d->op1; - d->op1 = x; - } - - if (TARGET_SIMD) - { - if (aarch64_evpc_zip (d)) - return true; - else if (aarch64_evpc_uzp (d)) - return true; - else if (aarch64_evpc_trn (d)) - return true; - return aarch64_evpc_tbl (d); - } - return false; -} - -/* Expand a vec_perm_const pattern. */ - -bool -aarch64_expand_vec_perm_const (rtx target, rtx op0, rtx op1, rtx sel) -{ - struct expand_vec_perm_d d; - int i, nelt, which; - - d.target = target; - d.op0 = op0; - d.op1 = op1; - - d.vmode = GET_MODE (target); - gcc_assert (VECTOR_MODE_P (d.vmode)); - d.nelt = nelt = GET_MODE_NUNITS (d.vmode); - d.testing_p = false; - - for (i = which = 0; i < nelt; ++i) - { - rtx e = XVECEXP (sel, 0, i); - int ei = INTVAL (e) & (2 * nelt - 1); - which |= (ei < nelt ? 1 : 2); - d.perm[i] = ei; - } - - switch (which) - { - default: - gcc_unreachable (); - - case 3: - d.one_vector_p = false; - if (!rtx_equal_p (op0, op1)) - break; - - /* The elements of PERM do not suggest that only the first operand - is used, but both operands are identical. Allow easier matching - of the permutation by folding the permutation into the single - input vector. */ - /* Fall Through. */ - case 2: - for (i = 0; i < nelt; ++i) - d.perm[i] &= nelt - 1; - d.op0 = op1; - d.one_vector_p = true; - break; - - case 1: - d.op1 = op0; - d.one_vector_p = true; - break; - } - - return aarch64_expand_vec_perm_const_1 (&d); -} - -static bool -aarch64_vectorize_vec_perm_const_ok (enum machine_mode vmode, - const unsigned char *sel) -{ - struct expand_vec_perm_d d; - unsigned int i, nelt, which; - bool ret; - - d.vmode = vmode; - d.nelt = nelt = GET_MODE_NUNITS (d.vmode); - d.testing_p = true; - memcpy (d.perm, sel, nelt); - - /* Calculate whether all elements are in one vector. */ - for (i = which = 0; i < nelt; ++i) - { - unsigned char e = d.perm[i]; - gcc_assert (e < 2 * nelt); - which |= (e < nelt ? 1 : 2); - } - - /* If all elements are from the second vector, reindex as if from the - first vector. */ - if (which == 2) - for (i = 0; i < nelt; ++i) - d.perm[i] -= nelt; - - /* Check whether the mask can be applied to a single vector. */ - d.one_vector_p = (which != 3); - - d.target = gen_raw_REG (d.vmode, LAST_VIRTUAL_REGISTER + 1); - d.op1 = d.op0 = gen_raw_REG (d.vmode, LAST_VIRTUAL_REGISTER + 2); - if (!d.one_vector_p) - d.op1 = gen_raw_REG (d.vmode, LAST_VIRTUAL_REGISTER + 3); - - start_sequence (); - ret = aarch64_expand_vec_perm_const_1 (&d); - end_sequence (); - - return ret; -} - -#undef TARGET_ADDRESS_COST -#define TARGET_ADDRESS_COST aarch64_address_cost - -/* This hook will determines whether unnamed bitfields affect the alignment - of the containing structure. The hook returns true if the structure - should inherit the alignment requirements of an unnamed bitfield's - type. */ -#undef TARGET_ALIGN_ANON_BITFIELD -#define TARGET_ALIGN_ANON_BITFIELD hook_bool_void_true - -#undef TARGET_ASM_ALIGNED_DI_OP -#define TARGET_ASM_ALIGNED_DI_OP "\t.xword\t" - -#undef TARGET_ASM_ALIGNED_HI_OP -#define TARGET_ASM_ALIGNED_HI_OP "\t.hword\t" - -#undef TARGET_ASM_ALIGNED_SI_OP -#define TARGET_ASM_ALIGNED_SI_OP "\t.word\t" - -#undef TARGET_ASM_CAN_OUTPUT_MI_THUNK -#define TARGET_ASM_CAN_OUTPUT_MI_THUNK \ - hook_bool_const_tree_hwi_hwi_const_tree_true - -#undef TARGET_ASM_FILE_START -#define TARGET_ASM_FILE_START aarch64_start_file - -#undef TARGET_ASM_OUTPUT_MI_THUNK -#define TARGET_ASM_OUTPUT_MI_THUNK aarch64_output_mi_thunk - -#undef TARGET_ASM_SELECT_RTX_SECTION -#define TARGET_ASM_SELECT_RTX_SECTION aarch64_select_rtx_section - -#undef TARGET_ASM_TRAMPOLINE_TEMPLATE -#define TARGET_ASM_TRAMPOLINE_TEMPLATE aarch64_asm_trampoline_template - -#undef TARGET_BUILD_BUILTIN_VA_LIST -#define TARGET_BUILD_BUILTIN_VA_LIST aarch64_build_builtin_va_list - -#undef TARGET_CALLEE_COPIES -#define TARGET_CALLEE_COPIES hook_bool_CUMULATIVE_ARGS_mode_tree_bool_false - -#undef TARGET_CAN_ELIMINATE -#define TARGET_CAN_ELIMINATE aarch64_can_eliminate - -#undef TARGET_CANNOT_FORCE_CONST_MEM -#define TARGET_CANNOT_FORCE_CONST_MEM aarch64_cannot_force_const_mem - -#undef TARGET_CONDITIONAL_REGISTER_USAGE -#define TARGET_CONDITIONAL_REGISTER_USAGE aarch64_conditional_register_usage - -/* Only the least significant bit is used for initialization guard - variables. */ -#undef TARGET_CXX_GUARD_MASK_BIT -#define TARGET_CXX_GUARD_MASK_BIT hook_bool_void_true - -#undef TARGET_C_MODE_FOR_SUFFIX -#define TARGET_C_MODE_FOR_SUFFIX aarch64_c_mode_for_suffix - -#ifdef TARGET_BIG_ENDIAN_DEFAULT -#undef TARGET_DEFAULT_TARGET_FLAGS -#define TARGET_DEFAULT_TARGET_FLAGS (MASK_BIG_END) -#endif - -#undef TARGET_CLASS_MAX_NREGS -#define TARGET_CLASS_MAX_NREGS aarch64_class_max_nregs - -#undef TARGET_BUILTIN_DECL -#define TARGET_BUILTIN_DECL aarch64_builtin_decl - -#undef TARGET_EXPAND_BUILTIN -#define TARGET_EXPAND_BUILTIN aarch64_expand_builtin - -#undef TARGET_EXPAND_BUILTIN_VA_START -#define TARGET_EXPAND_BUILTIN_VA_START aarch64_expand_builtin_va_start - -#undef TARGET_FUNCTION_ARG -#define TARGET_FUNCTION_ARG aarch64_function_arg - -#undef TARGET_FUNCTION_ARG_ADVANCE -#define TARGET_FUNCTION_ARG_ADVANCE aarch64_function_arg_advance - -#undef TARGET_FUNCTION_ARG_BOUNDARY -#define TARGET_FUNCTION_ARG_BOUNDARY aarch64_function_arg_boundary - -#undef TARGET_FUNCTION_OK_FOR_SIBCALL -#define TARGET_FUNCTION_OK_FOR_SIBCALL aarch64_function_ok_for_sibcall - -#undef TARGET_FUNCTION_VALUE -#define TARGET_FUNCTION_VALUE aarch64_function_value - -#undef TARGET_FUNCTION_VALUE_REGNO_P -#define TARGET_FUNCTION_VALUE_REGNO_P aarch64_function_value_regno_p - -#undef TARGET_FRAME_POINTER_REQUIRED -#define TARGET_FRAME_POINTER_REQUIRED aarch64_frame_pointer_required - -#undef TARGET_GIMPLIFY_VA_ARG_EXPR -#define TARGET_GIMPLIFY_VA_ARG_EXPR aarch64_gimplify_va_arg_expr - -#undef TARGET_INIT_BUILTINS -#define TARGET_INIT_BUILTINS aarch64_init_builtins - -#undef TARGET_LEGITIMATE_ADDRESS_P -#define TARGET_LEGITIMATE_ADDRESS_P aarch64_legitimate_address_hook_p - -#undef TARGET_LEGITIMATE_CONSTANT_P -#define TARGET_LEGITIMATE_CONSTANT_P aarch64_legitimate_constant_p - -#undef TARGET_LIBGCC_CMP_RETURN_MODE -#define TARGET_LIBGCC_CMP_RETURN_MODE aarch64_libgcc_cmp_return_mode - -#undef TARGET_MANGLE_TYPE -#define TARGET_MANGLE_TYPE aarch64_mangle_type - -#undef TARGET_MEMORY_MOVE_COST -#define TARGET_MEMORY_MOVE_COST aarch64_memory_move_cost - -#undef TARGET_MUST_PASS_IN_STACK -#define TARGET_MUST_PASS_IN_STACK must_pass_in_stack_var_size - -/* This target hook should return true if accesses to volatile bitfields - should use the narrowest mode possible. It should return false if these - accesses should use the bitfield container type. */ -#undef TARGET_NARROW_VOLATILE_BITFIELD -#define TARGET_NARROW_VOLATILE_BITFIELD hook_bool_void_false - -#undef TARGET_OPTION_OVERRIDE -#define TARGET_OPTION_OVERRIDE aarch64_override_options - -#undef TARGET_OVERRIDE_OPTIONS_AFTER_CHANGE -#define TARGET_OVERRIDE_OPTIONS_AFTER_CHANGE \ - aarch64_override_options_after_change - -#undef TARGET_PASS_BY_REFERENCE -#define TARGET_PASS_BY_REFERENCE aarch64_pass_by_reference - -#undef TARGET_PREFERRED_RELOAD_CLASS -#define TARGET_PREFERRED_RELOAD_CLASS aarch64_preferred_reload_class - -#undef TARGET_SECONDARY_RELOAD -#define TARGET_SECONDARY_RELOAD aarch64_secondary_reload - -#undef TARGET_SHIFT_TRUNCATION_MASK -#define TARGET_SHIFT_TRUNCATION_MASK aarch64_shift_truncation_mask - -#undef TARGET_SETUP_INCOMING_VARARGS -#define TARGET_SETUP_INCOMING_VARARGS aarch64_setup_incoming_varargs - -#undef TARGET_STRUCT_VALUE_RTX -#define TARGET_STRUCT_VALUE_RTX aarch64_struct_value_rtx - -#undef TARGET_REGISTER_MOVE_COST -#define TARGET_REGISTER_MOVE_COST aarch64_register_move_cost - -#undef TARGET_RETURN_IN_MEMORY -#define TARGET_RETURN_IN_MEMORY aarch64_return_in_memory - -#undef TARGET_RETURN_IN_MSB -#define TARGET_RETURN_IN_MSB aarch64_return_in_msb - -#undef TARGET_RTX_COSTS -#define TARGET_RTX_COSTS aarch64_rtx_costs - -#undef TARGET_TRAMPOLINE_INIT -#define TARGET_TRAMPOLINE_INIT aarch64_trampoline_init - -#undef TARGET_USE_BLOCKS_FOR_CONSTANT_P -#define TARGET_USE_BLOCKS_FOR_CONSTANT_P aarch64_use_blocks_for_constant_p - -#undef TARGET_VECTOR_MODE_SUPPORTED_P -#define TARGET_VECTOR_MODE_SUPPORTED_P aarch64_vector_mode_supported_p - -#undef TARGET_ARRAY_MODE_SUPPORTED_P -#define TARGET_ARRAY_MODE_SUPPORTED_P aarch64_array_mode_supported_p - -#undef TARGET_VECTORIZE_PREFERRED_SIMD_MODE -#define TARGET_VECTORIZE_PREFERRED_SIMD_MODE aarch64_preferred_simd_mode - -#undef TARGET_VECTORIZE_BUILTINS -#define TARGET_VECTORIZE_BUILTINS - -#undef TARGET_VECTORIZE_BUILTIN_VECTORIZED_FUNCTION -#define TARGET_VECTORIZE_BUILTIN_VECTORIZED_FUNCTION \ - aarch64_builtin_vectorized_function - -#undef TARGET_VECTORIZE_AUTOVECTORIZE_VECTOR_SIZES -#define TARGET_VECTORIZE_AUTOVECTORIZE_VECTOR_SIZES \ - aarch64_autovectorize_vector_sizes - -/* Section anchor support. */ - -#undef TARGET_MIN_ANCHOR_OFFSET -#define TARGET_MIN_ANCHOR_OFFSET -256 - -/* Limit the maximum anchor offset to 4k-1, since that's the limit for a - byte offset; we can do much more for larger data types, but have no way - to determine the size of the access. We assume accesses are aligned. */ -#undef TARGET_MAX_ANCHOR_OFFSET -#define TARGET_MAX_ANCHOR_OFFSET 4095 - -#undef TARGET_VECTOR_ALIGNMENT -#define TARGET_VECTOR_ALIGNMENT aarch64_simd_vector_alignment - -#undef TARGET_VECTORIZE_VECTOR_ALIGNMENT_REACHABLE -#define TARGET_VECTORIZE_VECTOR_ALIGNMENT_REACHABLE \ - aarch64_simd_vector_alignment_reachable - -/* vec_perm support. */ - -#undef TARGET_VECTORIZE_VEC_PERM_CONST_OK -#define TARGET_VECTORIZE_VEC_PERM_CONST_OK \ - aarch64_vectorize_vec_perm_const_ok - - -#undef TARGET_FIXED_CONDITION_CODE_REGS -#define TARGET_FIXED_CONDITION_CODE_REGS aarch64_fixed_condition_code_regs - -struct gcc_target targetm = TARGET_INITIALIZER; - -#include "gt-aarch64.h" |