;; ARMv8-A crypto patterns. ;; Copyright (C) 2013-2020 Free Software Foundation, Inc. ;; Contributed by ARM Ltd. ;; This file is part of GCC. ;; GCC is free software; you can redistribute it and/or modify it ;; under the terms of the GNU General Public License as published ;; by the Free Software Foundation; either version 3, or (at your ;; option) any later version. ;; GCC is distributed in the hope that it will be useful, but WITHOUT ;; ANY WARRANTY; without even the implied warranty of MERCHANTABILITY ;; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public ;; License for more details. ;; You should have received a copy of the GNU General Public License ;; along with GCC; see the file COPYING3. If not see ;; . (define_insn "crypto_" [(set (match_operand: 0 "register_operand" "=w") (unspec: [(match_operand: 1 "register_operand" "w")] CRYPTO_AESMC))] "TARGET_CRYPTO" ".\\t%q0, %q1" [(set_attr "type" "")] ) (define_insn "crypto_" [(set (match_operand:V16QI 0 "register_operand" "=w") (unspec:V16QI [(xor:V16QI (match_operand:V16QI 1 "register_operand" "%0") (match_operand:V16QI 2 "register_operand" "w"))] CRYPTO_AES))] "TARGET_CRYPTO" ".\\t%q0, %q2" [(set_attr "type" "")] ) ;; When AESE/AESMC fusion is enabled we really want to keep the two together ;; and enforce the register dependency without scheduling or register ;; allocation messing up the order or introducing moves inbetween. ;; Mash the two together during combine. (define_insn "*aarch32_crypto_aese_fused" [(set (match_operand:V16QI 0 "register_operand" "=w") (unspec:V16QI [(unspec:V16QI [(xor:V16QI (match_operand:V16QI 1 "register_operand" "%0") (match_operand:V16QI 2 "register_operand" "w"))] UNSPEC_AESE)] UNSPEC_AESMC))] "TARGET_CRYPTO && arm_fusion_enabled_p (tune_params::FUSE_AES_AESMC)" "aese.8\\t%q0, %q2\;aesmc.8\\t%q0, %q0" [(set_attr "type" "crypto_aese") (set_attr "length" "8")] ) ;; When AESD/AESIMC fusion is enabled we really want to keep the two together ;; and enforce the register dependency without scheduling or register ;; allocation messing up the order or introducing moves inbetween. ;; Mash the two together during combine. (define_insn "*aarch32_crypto_aesd_fused" [(set (match_operand:V16QI 0 "register_operand" "=w") (unspec:V16QI [(unspec:V16QI [(xor:V16QI (match_operand:V16QI 1 "register_operand" "%0") (match_operand:V16QI 2 "register_operand" "w"))] UNSPEC_AESD)] UNSPEC_AESIMC))] "TARGET_CRYPTO && arm_fusion_enabled_p (tune_params::FUSE_AES_AESMC)" "aesd.8\\t%q0, %q2\;aesimc.8\\t%q0, %q0" [(set_attr "type" "crypto_aese") (set_attr "length" "8")] ) (define_insn "crypto_" [(set (match_operand: 0 "register_operand" "=w") (unspec: [(match_operand: 1 "register_operand" "0") (match_operand: 2 "register_operand" "w")] CRYPTO_BINARY))] "TARGET_CRYPTO" ".\\t%q0, %q2" [(set_attr "type" "")] ) (define_insn "crypto_" [(set (match_operand: 0 "register_operand" "=w") (unspec: [(match_operand: 1 "register_operand" "0") (match_operand: 2 "register_operand" "w") (match_operand: 3 "register_operand" "w")] CRYPTO_TERNARY))] "TARGET_CRYPTO" ".\\t%q0, %q2, %q3" [(set_attr "type" "")] ) /* The vec_select operation always selects index 0 from the lower V2SI subreg of the V4SI, adjusted for endianness. Required due to neon_vget_lane and neon_set_lane that change the element ordering in memory for big-endian. */ (define_expand "crypto_sha1h" [(set (match_operand:V4SI 0 "register_operand") (match_operand:V4SI 1 "register_operand"))] "TARGET_CRYPTO" { rtx op2 = GEN_INT (NEON_ENDIAN_LANE_N (V2SImode, 0)); emit_insn (gen_crypto_sha1h_lb (operands[0], operands[1], op2)); DONE; }) (define_insn "crypto_sha1h_lb" [(set (match_operand:V4SI 0 "register_operand" "=w") (unspec:V4SI [(vec_select:SI (match_operand:V4SI 1 "register_operand" "w") (parallel [(match_operand:SI 2 "immediate_operand" "i")]))] UNSPEC_SHA1H))] "TARGET_CRYPTO && INTVAL (operands[2]) == NEON_ENDIAN_LANE_N (V2SImode, 0)" "sha1h.32\\t%q0, %q1" [(set_attr "type" "crypto_sha1_fast")] ) (define_insn "crypto_vmullp64" [(set (match_operand:TI 0 "register_operand" "=w") (unspec:TI [(match_operand:DI 1 "register_operand" "w") (match_operand:DI 2 "register_operand" "w")] UNSPEC_VMULLP64))] "TARGET_CRYPTO" "vmull.p64\\t%q0, %P1, %P2" [(set_attr "type" "crypto_pmull")] ) /* The vec_select operation always selects index 0 from the lower V2SI subreg of the V4SI, adjusted for endianness. Required due to neon_vget_lane and neon_set_lane that change the element ordering in memory for big-endian. */ (define_expand "crypto_" [(set (match_operand:V4SI 0 "register_operand") (unspec: [(match_operand: 1 "register_operand") (match_operand: 2 "register_operand") (match_operand: 3 "register_operand")] CRYPTO_SELECTING))] "TARGET_CRYPTO" { rtx op4 = GEN_INT (NEON_ENDIAN_LANE_N (V2SImode, 0)); emit_insn (gen_crypto__lb (operands[0], operands[1], operands[2], operands[3], op4)); DONE; }) (define_insn "crypto__lb" [(set (match_operand:V4SI 0 "register_operand" "=w") (unspec: [(match_operand: 1 "register_operand" "0") (vec_select:SI (match_operand: 2 "register_operand" "w") (parallel [(match_operand:SI 4 "immediate_operand" "i")])) (match_operand: 3 "register_operand" "w")] CRYPTO_SELECTING))] "TARGET_CRYPTO && INTVAL (operands[4]) == NEON_ENDIAN_LANE_N (V2SImode, 0)" ".\\t%q0, %q2, %q3" [(set_attr "type" "")] )