;; ARM Cortex-A8 NEON scheduling description. ;; Copyright (C) 2007-2018 Free Software Foundation, Inc. ;; Contributed by CodeSourcery. ;; This file is part of GCC. ;; GCC is free software; you can redistribute it and/or modify it ;; under the terms of the GNU General Public License as published ;; by the Free Software Foundation; either version 3, or (at your ;; option) any later version. ;; GCC is distributed in the hope that it will be useful, but WITHOUT ;; ANY WARRANTY; without even the implied warranty of MERCHANTABILITY ;; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public ;; License for more details. ;; You should have received a copy of the GNU General Public License ;; along with GCC; see the file COPYING3. If not see ;; . (define_attr "cortex_a8_neon_type" "neon_int_1,neon_int_2,neon_int_3,neon_int_4,neon_int_5,neon_vqneg_vqabs, neon_bit_ops_q, neon_vaba,neon_vaba_qqq, neon_vmov, neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,neon_mul_qqq_8_16_32_ddd_32, neon_mul_qdd_64_32_long_qqd_16_ddd_32_scalar_64_32_long_scalar, neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,neon_mla_qqq_8_16, neon_mla_ddd_32_qqd_16_ddd_32_scalar_qdd_64_32_long_scalar_qdd_64_32_long, neon_mla_qqq_32_qqd_32_scalar,neon_mul_ddd_16_scalar_32_16_long_scalar, neon_mul_qqd_32_scalar,neon_mla_ddd_16_scalar_qdd_32_16_long_scalar, neon_shift_1,neon_shift_2,neon_shift_3, neon_vqshl_vrshl_vqrshl_qqq,neon_vsra_vrsra,neon_fp_vadd_ddd_vabs_dd, neon_fp_vadd_qqq_vabs_qq,neon_fp_vsum,neon_fp_vmul_ddd,neon_fp_vmul_qqd, neon_fp_vmla_ddd,neon_fp_vmla_qqq,neon_fp_vmla_ddd_scalar, neon_fp_vmla_qqq_scalar,neon_fp_vrecps_vrsqrts_ddd, neon_fp_vrecps_vrsqrts_qqq,neon_bp_simple,neon_bp_2cycle,neon_bp_3cycle, neon_ldr,neon_str,neon_vld1_1_2_regs,neon_vld1_3_4_regs, neon_vld2_2_regs_vld1_vld2_all_lanes,neon_vld2_4_regs,neon_vld3_vld4, neon_vst1_1_2_regs_vst2_2_regs,neon_vst1_3_4_regs, neon_vst2_4_regs_vst3_vst4,neon_vld1_vld2_lane, neon_vld3_vld4_lane,neon_vst1_vst2_lane,neon_vst3_vst4_lane, neon_vld3_vld4_all_lanes,neon_mcr,neon_mcr_2_mcrr,neon_mrc,neon_mrrc, neon_ldm_2,neon_stm_2,none,unknown" (cond [ (eq_attr "type" "neon_logic, neon_logic_q,\ neon_bsl, neon_cls, neon_cnt,\ neon_add, neon_add_q") (const_string "neon_int_1") (eq_attr "type" "neon_add_widen, neon_sub_widen,\ neon_sub, neon_sub_q") (const_string "neon_int_2") (eq_attr "type" "neon_neg, neon_neg_q,\ neon_reduc_add, neon_reduc_add_q,\ neon_reduc_add_long,\ neon_add_long, neon_sub_long") (const_string "neon_int_3") (eq_attr "type" "neon_abs, neon_abs_q, neon_compare_zero, neon_compare_zero_q,\ neon_add_halve_narrow_q,\ neon_sub_halve_narrow_q,\ neon_add_halve, neon_add_halve_q,\ neon_qadd, neon_qadd_q,\ neon_tst, neon_tst_q") (const_string "neon_int_4") (eq_attr "type" "neon_abd_long, neon_sub_halve, neon_sub_halve_q,\ neon_qsub, neon_qsub_q,\ neon_abd, neon_abd_q,\ neon_compare, neon_compare_q,\ neon_minmax, neon_minmax_q, neon_reduc_minmax,\ neon_reduc_minmax_q") (const_string "neon_int_5") (eq_attr "type" "neon_qneg, neon_qneg_q, neon_qabs, neon_qabs_q") (const_string "neon_vqneg_vqabs") (eq_attr "type" "neon_move, neon_move_q") (const_string "neon_vmov") (eq_attr "type" "neon_bsl_q, neon_cls_q, neon_cnt_q") (const_string "neon_bit_ops_q") (eq_attr "type" "neon_arith_acc, neon_reduc_add_acc") (const_string "neon_vaba") (eq_attr "type" "neon_arith_acc_q") (const_string "neon_vaba_qqq") (eq_attr "type" "neon_shift_imm, neon_shift_imm_q,\ neon_shift_imm_long, neon_shift_imm_narrow_q,\ neon_shift_reg") (const_string "neon_shift_1") (eq_attr "type" "neon_sat_shift_imm, neon_sat_shift_imm_q, neon_sat_shift_imm_narrow_q,\ neon_sat_shift_reg") (const_string "neon_shift_2") (eq_attr "type" "neon_shift_reg_q") (const_string "neon_shift_3") (eq_attr "type" "neon_sat_shift_reg_q") (const_string "neon_vqshl_vrshl_vqrshl_qqq") (eq_attr "type" "neon_shift_acc, neon_shift_acc_q") (const_string "neon_vsra_vrsra") (eq_attr "type" "neon_mul_b, neon_mul_h,\ neon_mul_b_long, neon_mul_h_long,\ neon_sat_mul_b, neon_sat_mul_h,\ neon_sat_mul_b_long, neon_sat_mul_h_long") (const_string "neon_mul_ddd_8_16_qdd_16_8_long_32_16_long") (eq_attr "type" "neon_mul_b_q, neon_mul_h_q,\ neon_sat_mul_b_q, neon_sat_mul_h_q") (const_string "neon_mul_qqq_8_16_32_ddd_32") (eq_attr "type" "neon_mul_s, neon_mul_s_long,\ neon_sat_mul_s, neon_sat_mul_s_long,\ neon_mul_h_scalar_q, neon_sat_mul_h_scalar_q,\ neon_mul_s_scalar, neon_sat_mul_s_scalar,\ neon_mul_s_scalar_long,\ neon_sat_mul_s_scalar_long") (const_string "neon_mul_qdd_64_32_long_qqd_16_ddd_32_scalar_64_32_long_scalar") (eq_attr "type" "neon_mla_b, neon_mla_h,\ neon_mla_b_long, neon_mla_h_long,\ neon_sat_mla_b_long, neon_sat_mla_h_long,\ neon_sat_mla_h_scalar_long") (const_string "neon_mla_ddd_8_16_qdd_16_8_long_32_16_long") (eq_attr "type" "neon_mla_b_q, neon_mla_h_q") (const_string "neon_mla_qqq_8_16") (eq_attr "type" "neon_mla_s, neon_mla_s_long,\ neon_sat_mla_s_long,\ neon_mla_h_scalar_q, neon_mla_s_scalar,\ neon_mla_s_scalar_long,\ neon_sat_mla_s_scalar_long") (const_string "neon_mla_ddd_32_qqd_16_ddd_32_scalar_qdd_64_32_long_scalar_qdd_64_32_long") (eq_attr "type" "neon_mla_s_q, neon_mla_s_scalar_q") (const_string "neon_mla_qqq_32_qqd_32_scalar") (eq_attr "type" "neon_mul_h_scalar, neon_sat_mul_h_scalar,\ neon_mul_h_scalar_long,\ neon_sat_mul_h_scalar_long") (const_string "neon_mul_ddd_16_scalar_32_16_long_scalar") (eq_attr "type" "neon_mul_s_q, neon_sat_mul_s_q,\ neon_mul_s_scalar_q") (const_string "neon_mul_qqd_32_scalar") (eq_attr "type" "neon_mla_h_scalar, neon_mla_h_scalar_long") (const_string "neon_mla_ddd_16_scalar_qdd_32_16_long_scalar") (eq_attr "type" "neon_fp_abd_s, neon_fp_abs_s, neon_fp_neg_s,\ neon_fp_addsub_s, neon_fp_compare_s,\ neon_fp_minmax_s, neon_fp_mul_s,\ neon_fp_recpe_s, neon_fp_rsqrte_s,\ neon_fp_to_int_s, neon_int_to_fp_s") (const_string "neon_fp_vadd_ddd_vabs_dd") (eq_attr "type" "neon_fp_abd_s_q, neon_fp_abs_s_q,\ neon_fp_neg_s_q,\ neon_fp_addsub_s_q, neon_fp_compare_s_q,\ neon_fp_minmax_s_q, neon_fp_mul_s_q,\ neon_fp_recpe_s_q, neon_fp_rsqrte_s_q,\ neon_fp_to_int_s_q, neon_int_to_fp_s_q") (const_string "neon_fp_vadd_qqq_vabs_qq") (eq_attr "type" "neon_fp_reduc_add_s, neon_fp_reduc_minmax_s,\ neon_fp_reduc_add_s_q, neon_fp_reduc_minmax_s_q") (const_string "neon_fp_vsum") (eq_attr "type" "neon_fp_mul_s_scalar") (const_string "neon_fp_vmul_ddd") (eq_attr "type" "neon_fp_mul_s_scalar_q") (const_string "neon_fp_vmul_qqd") (eq_attr "type" "neon_fp_mla_s") (const_string "neon_fp_vmla_ddd") (eq_attr "type" "neon_fp_mla_s_q") (const_string "neon_fp_vmla_qqq") (eq_attr "type" "neon_fp_mla_s_scalar") (const_string "neon_fp_vmla_ddd_scalar") (eq_attr "type" "neon_fp_mla_s_scalar_q") (const_string "neon_fp_vmla_qqq_scalar") (eq_attr "type" "neon_fp_recps_s, neon_fp_rsqrts_s") (const_string "neon_fp_vrecps_vrsqrts_ddd") (eq_attr "type" "neon_fp_recps_s_q, neon_fp_rsqrts_s_q") (const_string "neon_fp_vrecps_vrsqrts_qqq") (eq_attr "type" "neon_move_narrow_q, neon_dup,\ neon_dup_q, neon_permute, neon_zip,\ neon_ext, neon_rev, neon_rev_q") (const_string "neon_bp_simple") (eq_attr "type" "neon_permute_q, neon_ext_q, neon_tbl1, neon_tbl2") (const_string "neon_bp_2cycle") (eq_attr "type" "neon_zip_q, neon_tbl3, neon_tbl4") (const_string "neon_bp_3cycle") (eq_attr "type" "neon_ldr") (const_string "neon_ldr") (eq_attr "type" "neon_str") (const_string "neon_str") (eq_attr "type" "neon_load1_1reg, neon_load1_1reg_q,\ neon_load1_2reg, neon_load1_2reg_q,\ neon_load2_2reg, neon_load2_2reg_q") (const_string "neon_vld1_1_2_regs") (eq_attr "type" "neon_load1_3reg, neon_load1_3reg_q,\ neon_load1_4reg, neon_load1_4reg_q") (const_string "neon_vld1_3_4_regs") (eq_attr "type" "neon_load1_all_lanes, neon_load1_all_lanes_q,\ neon_load2_all_lanes, neon_load2_all_lanes_q") (const_string "neon_vld2_2_regs_vld1_vld2_all_lanes") (eq_attr "type" "neon_load3_all_lanes, neon_load3_all_lanes_q,\ neon_load4_all_lanes, neon_load4_all_lanes_q,\ neon_load2_4reg, neon_load2_4reg_q") (const_string "neon_vld2_4_regs") (eq_attr "type" "neon_load3_3reg, neon_load3_3reg_q,\ neon_load4_4reg, neon_load4_4reg_q") (const_string "neon_vld3_vld4") (eq_attr "type" "f_loads, f_loadd, f_stores, f_stored,\ neon_load1_one_lane, neon_load1_one_lane_q,\ neon_load2_one_lane, neon_load2_one_lane_q") (const_string "neon_vld1_vld2_lane") (eq_attr "type" "neon_load3_one_lane, neon_load3_one_lane_q,\ neon_load4_one_lane, neon_load4_one_lane_q") (const_string "neon_vld3_vld4_lane") (eq_attr "type" "neon_store1_1reg, neon_store1_1reg_q,\ neon_store1_2reg, neon_store1_2reg_q,\ neon_store2_2reg, neon_store2_2reg_q") (const_string "neon_vst1_1_2_regs_vst2_2_regs") (eq_attr "type" "neon_store1_3reg, neon_store1_3reg_q,\ neon_store1_4reg, neon_store1_4reg_q") (const_string "neon_vst1_3_4_regs") (eq_attr "type" "neon_store2_4reg, neon_store2_4reg_q,\ neon_store3_3reg, neon_store3_3reg_q,\ neon_store4_4reg, neon_store4_4reg_q") (const_string "neon_vst2_4_regs_vst3_vst4") (eq_attr "type" "neon_store1_one_lane, neon_store1_one_lane_q,\ neon_store2_one_lane, neon_store2_one_lane_q") (const_string "neon_vst1_vst2_lane") (eq_attr "type" "neon_store3_one_lane, neon_store3_one_lane_q,\ neon_store4_one_lane, neon_store4_one_lane_q") (const_string "neon_vst3_vst4_lane") (eq_attr "type" "neon_from_gp, f_mcr") (const_string "neon_mcr") (eq_attr "type" "neon_from_gp_q, f_mcrr") (const_string "neon_mcr_2_mcrr") (eq_attr "type" "neon_to_gp, f_mrc") (const_string "neon_mrc") (eq_attr "type" "neon_to_gp_q, f_mrrc") (const_string "neon_mrrc")] (const_string "unknown"))) (define_automaton "cortex_a8_neon") ;; Only one load, store, permute, MCR or MRC instruction can be issued ;; per cycle. (define_cpu_unit "cortex_a8_neon_issue_perm" "cortex_a8_neon") ;; Only one data-processing instruction can be issued per cycle. (define_cpu_unit "cortex_a8_neon_issue_dp" "cortex_a8_neon") ;; The VFPLite unit (non-pipelined). (define_cpu_unit "cortex_a8_vfplite" "cortex_a8_neon") ;; We need a special mutual exclusion (to be used in addition to ;; cortex_a8_neon_issue_dp) for the case when an instruction such as ;; vmla.f is forwarded from E5 of the floating-point multiply pipeline to ;; E2 of the floating-point add pipeline. On the cycle previous to that ;; forward we must prevent issue of any instruction to the floating-point ;; add pipeline, but still allow issue of a data-processing instruction ;; to any of the other pipelines. (define_cpu_unit "cortex_a8_neon_issue_fadd" "cortex_a8_neon") ;; Patterns of reservation. ;; We model the NEON issue units as running in parallel with the core ones. ;; We assume that multi-cycle NEON instructions get decomposed into ;; micro-ops as they are issued into the NEON pipeline, and not as they ;; are issued into the ARM pipeline. Dual issue may not occur except ;; upon the first and last cycles of a multi-cycle instruction, but it ;; is unclear whether two multi-cycle instructions can issue together (in ;; this model they cannot). It is also unclear whether a pair of ;; a multi-cycle and single-cycle instructions, that could potentially ;; issue together, only do so if (say) the single-cycle one precedes ;; the other. (define_reservation "cortex_a8_neon_dp" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp") (define_reservation "cortex_a8_neon_dp_2" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp,\ cortex_a8_neon_issue_dp") (define_reservation "cortex_a8_neon_dp_4" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp") (define_reservation "cortex_a8_neon_fadd" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp+\ cortex_a8_neon_issue_fadd") (define_reservation "cortex_a8_neon_fadd_2" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp+\ cortex_a8_neon_issue_fadd,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_fadd") (define_reservation "cortex_a8_neon_perm" "(cortex_a8_alu0|cortex_a8_alu1)+\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_perm_2" "(cortex_a8_alu0|cortex_a8_alu1)+\ cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_perm_3" "(cortex_a8_alu0|cortex_a8_alu1)+\ cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_ls" "cortex_a8_issue_ls+cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_ls_2" "cortex_a8_issue_ls+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_ls_3" "cortex_a8_issue_ls+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_ls_4" "cortex_a8_issue_ls+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_ls_5" "cortex_a8_issue_ls+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_dp+cortex_a8_neon_issue_perm,\ cortex_a8_neon_issue_perm") (define_reservation "cortex_a8_neon_fmul_then_fadd" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp,\ nothing*3,\ cortex_a8_neon_issue_fadd") (define_reservation "cortex_a8_neon_fmul_then_fadd_2" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp,\ cortex_a8_neon_issue_dp,\ nothing*2,\ cortex_a8_neon_issue_fadd,\ cortex_a8_neon_issue_fadd") ;; VFP instructions can only be single-issued into the NEON pipeline. (define_reservation "cortex_a8_vfp" "(cortex_a8_alu0|cortex_a8_alu1)+cortex_a8_neon_issue_dp+\ cortex_a8_neon_issue_perm+cortex_a8_vfplite") ;; VFP instructions. ;; The VFPLite unit that executes these isn't pipelined; we give the ;; worst-case latencies (and choose the double-precision ones where we ;; do not distinguish on precision). We assume RunFast mode is not ;; enabled and therefore do not model the possible VFP instruction ;; execution in the NEON floating point pipelines, nor additional ;; latencies for the processing of subnormals. ;; ;; TODO: RunFast mode could potentially be enabled when -ffast-math ;; is specified. (define_insn_reservation "cortex_a8_vfp_add_sub" 10 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fconsts,fconstd,fadds,faddd")) "cortex_a8_vfp,cortex_a8_vfplite*9") (define_insn_reservation "cortex_a8_vfp_muls" 12 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fmuls")) "cortex_a8_vfp,cortex_a8_vfplite*11") ;; Don't model a reservation for more than 15 cycles as this explodes the ;; state space of the automaton for little gain. It is unlikely that the ;; scheduler will find enough instructions to hide the full latency of the ;; instructions. (define_insn_reservation "cortex_a8_vfp_muld" 17 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fmuld")) "cortex_a8_vfp,cortex_a8_vfplite*15") (define_insn_reservation "cortex_a8_vfp_macs" 21 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fmacs,ffmas")) "cortex_a8_vfp,cortex_a8_vfplite*15") (define_insn_reservation "cortex_a8_vfp_macd" 26 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fmacd,ffmad")) "cortex_a8_vfp,cortex_a8_vfplite*15") (define_insn_reservation "cortex_a8_vfp_divs" 37 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fdivs, fsqrts")) "cortex_a8_vfp,cortex_a8_vfplite*15") (define_insn_reservation "cortex_a8_vfp_divd" 65 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fdivd, fsqrtd")) "cortex_a8_vfp,cortex_a8_vfplite*15") ;; Comparisons can actually take 7 cycles sometimes instead of four, ;; but given all the other instructions lumped into type=ffarith that ;; take four cycles, we pick that latency. (define_insn_reservation "cortex_a8_vfp_farith" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "fmov,ffariths,ffarithd,fconsts,fconstd,fcmps,fcmpd")) "cortex_a8_vfp,cortex_a8_vfplite*3") (define_insn_reservation "cortex_a8_vfp_cvt" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "f_cvt,f_cvtf2i,f_cvti2f")) "cortex_a8_vfp,cortex_a8_vfplite*6") ;; NEON -> core transfers. (define_insn_reservation "cortex_a8_neon_mrc" 20 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mrc")) "cortex_a8_neon_ls") (define_insn_reservation "cortex_a8_neon_mrrc" 21 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mrrc")) "cortex_a8_neon_ls_2") ;; Arithmetic Operations ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N3. (define_insn_reservation "cortex_a8_neon_int_1" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_int_1")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)m operands at N1, ;; their (D|Q)n operands at N2, and produce a result at N3. (define_insn_reservation "cortex_a8_neon_int_2" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_int_2")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N3. (define_insn_reservation "cortex_a8_neon_int_3" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_int_3")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N4. (define_insn_reservation "cortex_a8_neon_int_4" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_int_4")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)m operands at N1, ;; their (D|Q)n operands at N2, and produce a result at N4. (define_insn_reservation "cortex_a8_neon_int_5" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_int_5")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N4. (define_insn_reservation "cortex_a8_neon_vqneg_vqabs" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vqneg_vqabs")) "cortex_a8_neon_dp") ;; Instructions using this reservation produce a result at N3. (define_insn_reservation "cortex_a8_neon_vmov" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vmov")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N6. (define_insn_reservation "cortex_a8_neon_vaba" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vaba")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N6 on cycle 2. (define_insn_reservation "cortex_a8_neon_vaba_qqq" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vaba_qqq")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N3 on cycle 2. (define_insn_reservation "cortex_a8_neon_bit_ops_q" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_bit_ops_q")) "cortex_a8_neon_dp_2") ;; Integer Multiply/Accumulate Operations ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N6. (define_insn_reservation "cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mul_ddd_8_16_qdd_16_8_long_32_16_long")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N6 on cycle 2. (define_insn_reservation "cortex_a8_neon_mul_qqq_8_16_32_ddd_32" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mul_qqq_8_16_32_ddd_32")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, and produce a result at N6 on cycle 2. (define_insn_reservation "cortex_a8_neon_mul_qdd_64_32_long_qqd_16_ddd_32_scalar_64_32_long_scalar" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mul_qdd_64_32_long_qqd_16_ddd_32_scalar_64_32_long_scalar")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N2, their (D|Q)d operands at N3, and ;; produce a result at N6. (define_insn_reservation "cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mla_ddd_8_16_qdd_16_8_long_32_16_long")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N2, their (D|Q)d operands at N3, and ;; produce a result at N6 on cycle 2. (define_insn_reservation "cortex_a8_neon_mla_qqq_8_16" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mla_qqq_8_16")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N6 on cycle 2. (define_insn_reservation "cortex_a8_neon_mla_ddd_32_qqd_16_ddd_32_scalar_qdd_64_32_long_scalar_qdd_64_32_long" 7 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mla_ddd_32_qqd_16_ddd_32_scalar_qdd_64_32_long_scalar_qdd_64_32_long")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N6 on cycle 4. (define_insn_reservation "cortex_a8_neon_mla_qqq_32_qqd_32_scalar" 9 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mla_qqq_32_qqd_32_scalar")) "cortex_a8_neon_dp_4") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, and produce a result at N6. (define_insn_reservation "cortex_a8_neon_mul_ddd_16_scalar_32_16_long_scalar" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mul_ddd_16_scalar_32_16_long_scalar")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, and produce a result at N6 on cycle 4. (define_insn_reservation "cortex_a8_neon_mul_qqd_32_scalar" 9 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mul_qqd_32_scalar")) "cortex_a8_neon_dp_4") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N6. (define_insn_reservation "cortex_a8_neon_mla_ddd_16_scalar_qdd_32_16_long_scalar" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mla_ddd_16_scalar_qdd_32_16_long_scalar")) "cortex_a8_neon_dp") ;; Shift Operations ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N3. (define_insn_reservation "cortex_a8_neon_shift_1" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_shift_1")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N4. (define_insn_reservation "cortex_a8_neon_shift_2" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_shift_2")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N3 on cycle 2. (define_insn_reservation "cortex_a8_neon_shift_3" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_shift_3")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N4 on cycle 2. (define_insn_reservation "cortex_a8_neon_vqshl_vrshl_vqrshl_qqq" 5 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vqshl_vrshl_vqrshl_qqq")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)m operands at N1, ;; their (D|Q)d operands at N3, and produce a result at N6. (define_insn_reservation "cortex_a8_neon_vsra_vrsra" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vsra_vrsra")) "cortex_a8_neon_dp") ;; Floating point Operations ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N5. (define_insn_reservation "cortex_a8_neon_fp_vadd_ddd_vabs_dd" 5 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vadd_ddd_vabs_dd")) "cortex_a8_neon_fadd") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N5 on cycle 2. (define_insn_reservation "cortex_a8_neon_fp_vadd_qqq_vabs_qq" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vadd_qqq_vabs_qq")) "cortex_a8_neon_fadd_2") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N5. (define_insn_reservation "cortex_a8_neon_fp_vsum" 5 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vsum")) "cortex_a8_neon_fadd") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, and produce a result at N5. (define_insn_reservation "cortex_a8_neon_fp_vmul_ddd" 5 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmul_ddd")) "cortex_a8_neon_dp") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, and produce a result at N5 on cycle 2. (define_insn_reservation "cortex_a8_neon_fp_vmul_qqd" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmul_qqd")) "cortex_a8_neon_dp_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N2, their (D|Q)d operands at N3, and ;; produce a result at N9. (define_insn_reservation "cortex_a8_neon_fp_vmla_ddd" 9 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmla_ddd")) "cortex_a8_neon_fmul_then_fadd") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N2, their (D|Q)d operands at N3, and ;; produce a result at N9 on cycle 2. (define_insn_reservation "cortex_a8_neon_fp_vmla_qqq" 10 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmla_qqq")) "cortex_a8_neon_fmul_then_fadd_2") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N9. (define_insn_reservation "cortex_a8_neon_fp_vmla_ddd_scalar" 9 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmla_ddd_scalar")) "cortex_a8_neon_fmul_then_fadd") ;; Instructions using this reservation read their (D|Q)n operands at N2, ;; their (D|Q)m operands at N1, their (D|Q)d operands at N3, and ;; produce a result at N9 on cycle 2. (define_insn_reservation "cortex_a8_neon_fp_vmla_qqq_scalar" 10 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vmla_qqq_scalar")) "cortex_a8_neon_fmul_then_fadd_2") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N9. (define_insn_reservation "cortex_a8_neon_fp_vrecps_vrsqrts_ddd" 9 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_fp_vrecps_vrsqrts_ddd")) "cortex_a8_neon_fmul_then_fadd") ;; Instructions using this reservation read their source operands at N2, and ;; produce a result at N9 on cycle 2. (define_insn_reservation "cortex_a8_neon_fp_vrecps_vrsqrts_qqq" 10 (and (eq_attr "tune" "cortexa8") (eq_attr "type" "neon_fp_recps_s_q, neon_fp_rsqrts_s_q")) "cortex_a8_neon_fmul_then_fadd_2") ;; Permute operations. ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N2. (define_insn_reservation "cortex_a8_neon_bp_simple" 2 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_bp_simple")) "cortex_a8_neon_perm") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N2 on cycle 2. (define_insn_reservation "cortex_a8_neon_bp_2cycle" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_bp_2cycle")) "cortex_a8_neon_perm_2") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N2 on cycle 3. (define_insn_reservation "cortex_a8_neon_bp_3cycle" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_bp_3cycle")) "cortex_a8_neon_perm_3") ;; Load Operations. ;; Instructions using this reservation produce a result at N1. (define_insn_reservation "cortex_a8_neon_ldr" 1 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_ldr")) "cortex_a8_neon_ls") ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_str" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_str")) "cortex_a8_neon_ls") ;; Instructions using this reservation produce a result at N1 on cycle 2. (define_insn_reservation "cortex_a8_neon_vld1_1_2_regs" 2 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld1_1_2_regs")) "cortex_a8_neon_ls_2") ;; Instructions using this reservation produce a result at N1 on cycle 3. (define_insn_reservation "cortex_a8_neon_vld1_3_4_regs" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld1_3_4_regs")) "cortex_a8_neon_ls_3") ;; Instructions using this reservation produce a result at N2 on cycle 2. (define_insn_reservation "cortex_a8_neon_vld2_2_regs_vld1_vld2_all_lanes" 3 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld2_2_regs_vld1_vld2_all_lanes")) "cortex_a8_neon_ls_2") ;; Instructions using this reservation produce a result at N2 on cycle 3. (define_insn_reservation "cortex_a8_neon_vld2_4_regs" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld2_4_regs")) "cortex_a8_neon_ls_3") ;; Instructions using this reservation produce a result at N2 on cycle 4. (define_insn_reservation "cortex_a8_neon_vld3_vld4" 5 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld3_vld4")) "cortex_a8_neon_ls_4") ;; Store operations. ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_vst1_1_2_regs_vst2_2_regs" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vst1_1_2_regs_vst2_2_regs")) "cortex_a8_neon_ls_2") ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_vst1_3_4_regs" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vst1_3_4_regs")) "cortex_a8_neon_ls_3") ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_vst2_4_regs_vst3_vst4" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vst2_4_regs_vst3_vst4")) "cortex_a8_neon_ls_4") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N2 on cycle 3. (define_insn_reservation "cortex_a8_neon_vld1_vld2_lane" 4 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld1_vld2_lane")) "cortex_a8_neon_ls_3") ;; Instructions using this reservation read their source operands at N1, and ;; produce a result at N2 on cycle 5. (define_insn_reservation "cortex_a8_neon_vld3_vld4_lane" 6 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vld3_vld4_lane")) "cortex_a8_neon_ls_5") ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_vst1_vst2_lane" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vst1_vst2_lane")) "cortex_a8_neon_ls_2") ;; Instructions using this reservation read their source operands at N1. (define_insn_reservation "cortex_a8_neon_vst3_vst4_lane" 0 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_vst3_vst4_lane")) "cortex_a8_neon_ls_3") ;; Register Transfer Operations ;; Instructions using this reservation produce a result at N2. (define_insn_reservation "cortex_a8_neon_mcr" 2 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mcr")) "cortex_a8_neon_perm") ;; Instructions using this reservation produce a result at N2. (define_insn_reservation "cortex_a8_neon_mcr_2_mcrr" 2 (and (eq_attr "tune" "cortexa8") (eq_attr "cortex_a8_neon_type" "neon_mcr_2_mcrr")) "cortex_a8_neon_perm_2") ;; Exceptions to the default latencies. (define_bypass 1 "cortex_a8_neon_mcr_2_mcrr" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 1 "cortex_a8_neon_mcr" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_vld3_vld4_lane" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_vld1_vld2_lane" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 4 "cortex_a8_neon_vld3_vld4" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_vld2_4_regs" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_vld2_2_regs_vld1_vld2_all_lanes" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_vld1_3_4_regs" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 1 "cortex_a8_neon_vld1_1_2_regs" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 0 "cortex_a8_neon_ldr" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_bp_3cycle" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_bp_2cycle" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 1 "cortex_a8_neon_bp_simple" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 9 "cortex_a8_neon_fp_vrecps_vrsqrts_qqq" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 8 "cortex_a8_neon_fp_vrecps_vrsqrts_ddd" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 9 "cortex_a8_neon_fp_vmla_qqq_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 8 "cortex_a8_neon_fp_vmla_ddd_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 9 "cortex_a8_neon_fp_vmla_qqq" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 8 "cortex_a8_neon_fp_vmla_ddd" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_fp_vmul_qqd" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 4 "cortex_a8_neon_fp_vmul_ddd" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 4 "cortex_a8_neon_fp_vsum" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_fp_vadd_qqq_vabs_qq" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 4 "cortex_a8_neon_fp_vadd_ddd_vabs_dd" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_vsra_vrsra" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 4 "cortex_a8_neon_vqshl_vrshl_vqrshl_qqq" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_shift_3" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_shift_2" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_shift_1" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_mla_ddd_16_scalar_qdd_32_16_long_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 8 "cortex_a8_neon_mul_qqd_32_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_mul_ddd_16_scalar_32_16_long_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 8 "cortex_a8_neon_mla_qqq_32_qqd_32_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 6 "cortex_a8_neon_mla_ddd_32_qqd_16_ddd_32_scalar_qdd_64_32_long_scalar_qdd_64_32_long" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 6 "cortex_a8_neon_mla_qqq_8_16" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 6 "cortex_a8_neon_mul_qdd_64_32_long_qqd_16_ddd_32_scalar_64_32_long_scalar" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 6 "cortex_a8_neon_mul_qqq_8_16_32_ddd_32" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 6 "cortex_a8_neon_vaba_qqq" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 5 "cortex_a8_neon_vaba" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_bit_ops_q" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_vqneg_vqabs" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_int_5" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 3 "cortex_a8_neon_int_4" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_int_3" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_int_2" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq") (define_bypass 2 "cortex_a8_neon_int_1" "cortex_a8_neon_int_1,\ cortex_a8_neon_int_4,\ cortex_a8_neon_bit_ops_q,\ cortex_a8_neon_mul_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mul_qqq_8_16_32_ddd_32,\ cortex_a8_neon_mla_ddd_8_16_qdd_16_8_long_32_16_long,\ cortex_a8_neon_mla_qqq_8_16,\ cortex_a8_neon_fp_vadd_ddd_vabs_dd,\ cortex_a8_neon_fp_vadd_qqq_vabs_qq,\ cortex_a8_neon_fp_vmla_ddd,\ cortex_a8_neon_fp_vmla_qqq,\ cortex_a8_neon_fp_vrecps_vrsqrts_ddd,\ cortex_a8_neon_fp_vrecps_vrsqrts_qqq")