;; VSX patterns. ;; Copyright (C) 2009, 2010, 2011, 2012 ;; Free Software Foundation, Inc. ;; Contributed by Michael Meissner ;; This file is part of GCC. ;; GCC is free software; you can redistribute it and/or modify it ;; under the terms of the GNU General Public License as published ;; by the Free Software Foundation; either version 3, or (at your ;; option) any later version. ;; GCC is distributed in the hope that it will be useful, but WITHOUT ;; ANY WARRANTY; without even the implied warranty of MERCHANTABILITY ;; or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public ;; License for more details. ;; You should have received a copy of the GNU General Public License ;; along with GCC; see the file COPYING3. If not see ;; . ;; Iterator for both scalar and vector floating point types supported by VSX (define_mode_iterator VSX_B [DF V4SF V2DF]) ;; Iterator for the 2 64-bit vector types (define_mode_iterator VSX_D [V2DF V2DI]) ;; Iterator for the 2 32-bit vector types (define_mode_iterator VSX_W [V4SF V4SI]) ;; Iterator for the DF types (define_mode_iterator VSX_DF [V2DF DF]) ;; Iterator for vector floating point types supported by VSX (define_mode_iterator VSX_F [V4SF V2DF]) ;; Iterator for logical types supported by VSX (define_mode_iterator VSX_L [V16QI V8HI V4SI V2DI V4SF V2DF TI]) ;; Iterator for memory move. Handle TImode specially to allow ;; it to use gprs as well as vsx registers. (define_mode_iterator VSX_M [V16QI V8HI V4SI V2DI V4SF V2DF]) ;; Map into the appropriate load/store name based on the type (define_mode_attr VSm [(V16QI "vw4") (V8HI "vw4") (V4SI "vw4") (V4SF "vw4") (V2DF "vd2") (V2DI "vd2") (DF "d") (TI "vw4")]) ;; Map into the appropriate suffix based on the type (define_mode_attr VSs [(V16QI "sp") (V8HI "sp") (V4SI "sp") (V4SF "sp") (V2DF "dp") (V2DI "dp") (DF "dp") (SF "sp") (TI "sp")]) ;; Map the register class used (define_mode_attr VSr [(V16QI "v") (V8HI "v") (V4SI "v") (V4SF "wf") (V2DI "wd") (V2DF "wd") (DF "ws") (SF "d") (TI "wd")]) ;; Map the register class used for float<->int conversions (define_mode_attr VSr2 [(V2DF "wd") (V4SF "wf") (DF "ws")]) (define_mode_attr VSr3 [(V2DF "wa") (V4SF "wa") (DF "ws")]) ;; Map the register class for sp<->dp float conversions, destination (define_mode_attr VSr4 [(SF "ws") (DF "f") (V2DF "wd") (V4SF "v")]) ;; Map the register class for sp<->dp float conversions, destination (define_mode_attr VSr5 [(SF "ws") (DF "f") (V2DF "v") (V4SF "wd")]) ;; Same size integer type for floating point data (define_mode_attr VSi [(V4SF "v4si") (V2DF "v2di") (DF "di")]) (define_mode_attr VSI [(V4SF "V4SI") (V2DF "V2DI") (DF "DI")]) ;; Word size for same size conversion (define_mode_attr VSc [(V4SF "w") (V2DF "d") (DF "d")]) ;; Map into either s or v, depending on whether this is a scalar or vector ;; operation (define_mode_attr VSv [(V16QI "v") (V8HI "v") (V4SI "v") (V4SF "v") (V2DI "v") (V2DF "v") (TI "v") (DF "s")]) ;; Appropriate type for add ops (and other simple FP ops) (define_mode_attr VStype_simple [(V2DF "vecdouble") (V4SF "vecfloat") (DF "fp")]) (define_mode_attr VSfptype_simple [(V2DF "fp_addsub_d") (V4SF "fp_addsub_s") (DF "fp_addsub_d")]) ;; Appropriate type for multiply ops (define_mode_attr VStype_mul [(V2DF "vecdouble") (V4SF "vecfloat") (DF "dmul")]) (define_mode_attr VSfptype_mul [(V2DF "fp_mul_d") (V4SF "fp_mul_s") (DF "fp_mul_d")]) ;; Appropriate type for divide ops. (define_mode_attr VStype_div [(V2DF "vecdiv") (V4SF "vecfdiv") (DF "ddiv")]) (define_mode_attr VSfptype_div [(V2DF "fp_div_d") (V4SF "fp_div_s") (DF "fp_div_d")]) ;; Appropriate type for sqrt ops. For now, just lump the vector sqrt with ;; the scalar sqrt (define_mode_attr VStype_sqrt [(V2DF "dsqrt") (V4SF "ssqrt") (DF "dsqrt")]) (define_mode_attr VSfptype_sqrt [(V2DF "fp_sqrt_d") (V4SF "fp_sqrt_s") (DF "fp_sqrt_d")]) ;; Iterator and modes for sp<->dp conversions ;; Because scalar SF values are represented internally as double, use the ;; V4SF type to represent this than SF. (define_mode_iterator VSX_SPDP [DF V4SF V2DF]) (define_mode_attr VS_spdp_res [(DF "V4SF") (V4SF "V2DF") (V2DF "V4SF")]) (define_mode_attr VS_spdp_insn [(DF "xscvdpsp") (V4SF "xvcvspdp") (V2DF "xvcvdpsp")]) (define_mode_attr VS_spdp_type [(DF "fp") (V4SF "vecdouble") (V2DF "vecdouble")]) ;; Map the scalar mode for a vector type (define_mode_attr VS_scalar [(V2DF "DF") (V2DI "DI") (V4SF "SF") (V4SI "SI") (V8HI "HI") (V16QI "QI")]) ;; Map to a double-sized vector mode (define_mode_attr VS_double [(V4SI "V8SI") (V4SF "V8SF") (V2DI "V4DI") (V2DF "V4DF")]) ;; Constants for creating unspecs (define_c_enum "unspec" [UNSPEC_VSX_CONCAT UNSPEC_VSX_CVDPSXWS UNSPEC_VSX_CVDPUXWS UNSPEC_VSX_CVSPDP UNSPEC_VSX_CVSXWDP UNSPEC_VSX_CVUXWDP UNSPEC_VSX_CVSXDSP UNSPEC_VSX_CVUXDSP UNSPEC_VSX_CVSPSXDS UNSPEC_VSX_CVSPUXDS UNSPEC_VSX_TDIV UNSPEC_VSX_TSQRT UNSPEC_VSX_SET UNSPEC_VSX_ROUND_I UNSPEC_VSX_ROUND_IC UNSPEC_VSX_SLDWI ]) ;; VSX moves (define_insn "*vsx_mov" [(set (match_operand:VSX_M 0 "nonimmediate_operand" "=Z,,,?Z,?wa,?wa,*o,*r,*r,,?wa,v,wZ,v") (match_operand:VSX_M 1 "input_operand" ",Z,,wa,Z,wa,r,o,r,j,j,W,v,wZ"))] "VECTOR_MEM_VSX_P (mode) && (register_operand (operands[0], mode) || register_operand (operands[1], mode))" { switch (which_alternative) { case 0: case 3: gcc_assert (MEM_P (operands[0]) && GET_CODE (XEXP (operands[0], 0)) != PRE_INC && GET_CODE (XEXP (operands[0], 0)) != PRE_DEC && GET_CODE (XEXP (operands[0], 0)) != PRE_MODIFY); return "stxx %x1,%y0"; case 1: case 4: gcc_assert (MEM_P (operands[1]) && GET_CODE (XEXP (operands[1], 0)) != PRE_INC && GET_CODE (XEXP (operands[1], 0)) != PRE_DEC && GET_CODE (XEXP (operands[1], 0)) != PRE_MODIFY); return "lxx %x0,%y1"; case 2: case 5: return "xxlor %x0,%x1,%x1"; case 6: case 7: case 8: return "#"; case 9: case 10: return "xxlxor %x0,%x0,%x0"; case 11: return output_vec_const_move (operands); case 12: gcc_assert (MEM_P (operands[0]) && GET_CODE (XEXP (operands[0], 0)) != PRE_INC && GET_CODE (XEXP (operands[0], 0)) != PRE_DEC && GET_CODE (XEXP (operands[0], 0)) != PRE_MODIFY); return "stvx %1,%y0"; case 13: gcc_assert (MEM_P (operands[0]) && GET_CODE (XEXP (operands[0], 0)) != PRE_INC && GET_CODE (XEXP (operands[0], 0)) != PRE_DEC && GET_CODE (XEXP (operands[0], 0)) != PRE_MODIFY); return "lvx %0,%y1"; default: gcc_unreachable (); } } [(set_attr "type" "vecstore,vecload,vecsimple,vecstore,vecload,vecsimple,*,*,*,vecsimple,vecsimple,*,vecstore,vecload")]) ;; Unlike other VSX moves, allow the GPRs, since a normal use of TImode is for ;; unions. However for plain data movement, slightly favor the vector loads (define_insn "*vsx_movti" [(set (match_operand:TI 0 "nonimmediate_operand" "=Z,wa,wa,?o,?r,?r,wa,v,v,wZ") (match_operand:TI 1 "input_operand" "wa,Z,wa,r,o,r,j,W,wZ,v"))] "VECTOR_MEM_VSX_P (TImode) && (register_operand (operands[0], TImode) || register_operand (operands[1], TImode))" { switch (which_alternative) { case 0: return "stxvd2x %x1,%y0"; case 1: return "lxvd2x %x0,%y1"; case 2: return "xxlor %x0,%x1,%x1"; case 3: case 4: case 5: return "#"; case 6: return "xxlxor %x0,%x0,%x0"; case 7: return output_vec_const_move (operands); case 8: return "stvx %1,%y0"; case 9: return "lvx %0,%y1"; default: gcc_unreachable (); } } [(set_attr "type" "vecstore,vecload,vecsimple,*,*,*,vecsimple,*,vecstore,vecload")]) ;; Explicit load/store expanders for the builtin functions (define_expand "vsx_load_" [(set (match_operand:VSX_M 0 "vsx_register_operand" "") (match_operand:VSX_M 1 "memory_operand" ""))] "VECTOR_MEM_VSX_P (mode)" "") (define_expand "vsx_store_" [(set (match_operand:VEC_M 0 "memory_operand" "") (match_operand:VEC_M 1 "vsx_register_operand" ""))] "VECTOR_MEM_VSX_P (mode)" "") ;; VSX scalar and vector floating point arithmetic instructions (define_insn "*vsx_add3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (plus:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xadd %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_sub3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (minus:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xsub %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_mul3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (mult:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xmul %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_div3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (div:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xdiv %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; *tdiv* instruction returning the FG flag (define_expand "vsx_tdiv3_fg" [(set (match_dup 3) (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" "") (match_operand:VSX_B 2 "vsx_register_operand" "")] UNSPEC_VSX_TDIV)) (set (match_operand:SI 0 "gpc_reg_operand" "") (gt:SI (match_dup 3) (const_int 0)))] "VECTOR_UNIT_VSX_P (mode)" { operands[3] = gen_reg_rtx (CCFPmode); }) ;; *tdiv* instruction returning the FE flag (define_expand "vsx_tdiv3_fe" [(set (match_dup 3) (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" "") (match_operand:VSX_B 2 "vsx_register_operand" "")] UNSPEC_VSX_TDIV)) (set (match_operand:SI 0 "gpc_reg_operand" "") (eq:SI (match_dup 3) (const_int 0)))] "VECTOR_UNIT_VSX_P (mode)" { operands[3] = gen_reg_rtx (CCFPmode); }) (define_insn "*vsx_tdiv3_internal" [(set (match_operand:CCFP 0 "cc_reg_operand" "=x,x") (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")] UNSPEC_VSX_TDIV))] "VECTOR_UNIT_VSX_P (mode)" "xtdiv %0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_fre2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_FRES))] "VECTOR_UNIT_VSX_P (mode)" "xre %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_neg2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (neg:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xneg %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_abs2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (abs:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xabs %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_nabs2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (neg:VSX_B (abs:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa"))))] "VECTOR_UNIT_VSX_P (mode)" "xnabs %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_smax3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (smax:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xmax %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_smin3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (smin:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xmin %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; Special VSX version of smin/smax for single precision floating point. Since ;; both numbers are rounded to single precision, we can just use the DP version ;; of the instruction. (define_insn "*vsx_smaxsf3" [(set (match_operand:SF 0 "vsx_register_operand" "=f") (smax:SF (match_operand:SF 1 "vsx_register_operand" "f") (match_operand:SF 2 "vsx_register_operand" "f")))] "VECTOR_UNIT_VSX_P (DFmode)" "xsmaxdp %x0,%x1,%x2" [(set_attr "type" "fp") (set_attr "fp_type" "fp_addsub_d")]) (define_insn "*vsx_sminsf3" [(set (match_operand:SF 0 "vsx_register_operand" "=f") (smin:SF (match_operand:SF 1 "vsx_register_operand" "f") (match_operand:SF 2 "vsx_register_operand" "f")))] "VECTOR_UNIT_VSX_P (DFmode)" "xsmindp %x0,%x1,%x2" [(set_attr "type" "fp") (set_attr "fp_type" "fp_addsub_d")]) (define_insn "*vsx_sqrt2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (sqrt:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xsqrt %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_rsqrte2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_RSQRT))] "VECTOR_UNIT_VSX_P (mode)" "xrsqrte %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; *tsqrt* returning the fg flag (define_expand "vsx_tsqrt2_fg" [(set (match_dup 3) (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" "")] UNSPEC_VSX_TSQRT)) (set (match_operand:SI 0 "gpc_reg_operand" "") (gt:SI (match_dup 3) (const_int 0)))] "VECTOR_UNIT_VSX_P (mode)" { operands[3] = gen_reg_rtx (CCFPmode); }) ;; *tsqrt* returning the fe flag (define_expand "vsx_tsqrt2_fe" [(set (match_dup 3) (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" "")] UNSPEC_VSX_TSQRT)) (set (match_operand:SI 0 "gpc_reg_operand" "") (eq:SI (match_dup 3) (const_int 0)))] "VECTOR_UNIT_VSX_P (mode)" { operands[3] = gen_reg_rtx (CCFPmode); }) (define_insn "*vsx_tsqrt2_internal" [(set (match_operand:CCFP 0 "cc_reg_operand" "=x,x") (unspec:CCFP [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_VSX_TSQRT))] "VECTOR_UNIT_VSX_P (mode)" "xtsqrt %0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; Fused vector multiply/add instructions Support the classical DF versions of ;; fma, which allows the target to be a separate register from the 3 inputs. ;; Under VSX, the target must be either the addend or the first multiply. ;; Where we can, also do the same for the Altivec V4SF fmas. (define_insn "*vsx_fmadf4" [(set (match_operand:DF 0 "vsx_register_operand" "=ws,ws,?wa,?wa,d") (fma:DF (match_operand:DF 1 "vsx_register_operand" "%ws,ws,wa,wa,d") (match_operand:DF 2 "vsx_register_operand" "ws,0,wa,0,d") (match_operand:DF 3 "vsx_register_operand" "0,ws,0,wa,d")))] "VECTOR_UNIT_VSX_P (DFmode)" "@ xsmaddadp %x0,%x1,%x2 xsmaddmdp %x0,%x1,%x3 xsmaddadp %x0,%x1,%x2 xsmaddmdp %x0,%x1,%x3 {fma|fmadd} %0,%1,%2,%3" [(set_attr "type" "fp") (set_attr "fp_type" "fp_maddsub_d")]) (define_insn "*vsx_fmav4sf4" [(set (match_operand:V4SF 0 "vsx_register_operand" "=ws,ws,?wa,?wa,v") (fma:V4SF (match_operand:V4SF 1 "vsx_register_operand" "%ws,ws,wa,wa,v") (match_operand:V4SF 2 "vsx_register_operand" "ws,0,wa,0,v") (match_operand:V4SF 3 "vsx_register_operand" "0,ws,0,wa,v")))] "VECTOR_UNIT_VSX_P (V4SFmode)" "@ xvmaddasp %x0,%x1,%x2 xvmaddmsp %x0,%x1,%x3 xvmaddasp %x0,%x1,%x2 xvmaddmsp %x0,%x1,%x3 vmaddfp %0,%1,%2,%3" [(set_attr "type" "vecfloat")]) (define_insn "*vsx_fmav2df4" [(set (match_operand:V2DF 0 "vsx_register_operand" "=ws,ws,?wa,?wa") (fma:V2DF (match_operand:V2DF 1 "vsx_register_operand" "%ws,ws,wa,wa") (match_operand:V2DF 2 "vsx_register_operand" "ws,0,wa,0") (match_operand:V2DF 3 "vsx_register_operand" "0,ws,0,wa")))] "VECTOR_UNIT_VSX_P (V2DFmode)" "@ xvmaddadp %x0,%x1,%x2 xvmaddmdp %x0,%x1,%x3 xvmaddadp %x0,%x1,%x2 xvmaddmdp %x0,%x1,%x3" [(set_attr "type" "vecdouble")]) (define_insn "*vsx_fmsdf4" [(set (match_operand:DF 0 "vsx_register_operand" "=ws,ws,?wa,?wa,d") (fma:DF (match_operand:DF 1 "vsx_register_operand" "%ws,ws,wa,wa,d") (match_operand:DF 2 "vsx_register_operand" "ws,0,wa,0,d") (neg:DF (match_operand:DF 3 "vsx_register_operand" "0,ws,0,wa,d"))))] "VECTOR_UNIT_VSX_P (DFmode)" "@ xsmsubadp %x0,%x1,%x2 xsmsubmdp %x0,%x1,%x3 xsmsubadp %x0,%x1,%x2 xsmsubmdp %x0,%x1,%x3 {fms|fmsub} %0,%1,%2,%3" [(set_attr "type" "fp") (set_attr "fp_type" "fp_maddsub_d")]) (define_insn "*vsx_fms4" [(set (match_operand:VSX_F 0 "vsx_register_operand" "=,,?wa,?wa") (fma:VSX_F (match_operand:VSX_F 1 "vsx_register_operand" "%,,wa,wa") (match_operand:VSX_F 2 "vsx_register_operand" ",0,wa,0") (neg:VSX_F (match_operand:VSX_F 3 "vsx_register_operand" "0,,0,wa"))))] "VECTOR_UNIT_VSX_P (mode)" "@ xmsuba %x0,%x1,%x2 xmsubm %x0,%x1,%x3 xmsuba %x0,%x1,%x2 xmsubm %x0,%x1,%x3" [(set_attr "type" "")]) (define_insn "*vsx_nfmadf4" [(set (match_operand:DF 0 "vsx_register_operand" "=ws,ws,?wa,?wa,d") (neg:DF (fma:DF (match_operand:DF 1 "vsx_register_operand" "ws,ws,wa,wa,d") (match_operand:DF 2 "vsx_register_operand" "ws,0,wa,0,d") (match_operand:DF 3 "vsx_register_operand" "0,ws,0,wa,d"))))] "VECTOR_UNIT_VSX_P (DFmode)" "@ xsnmaddadp %x0,%x1,%x2 xsnmaddmdp %x0,%x1,%x3 xsnmaddadp %x0,%x1,%x2 xsnmaddmdp %x0,%x1,%x3 {fnma|fnmadd} %0,%1,%2,%3" [(set_attr "type" "fp") (set_attr "fp_type" "fp_maddsub_d")]) (define_insn "*vsx_nfma4" [(set (match_operand:VSX_F 0 "vsx_register_operand" "=,,?wa,?wa") (neg:VSX_F (fma:VSX_F (match_operand:VSX_F 1 "vsx_register_operand" ",,wa,wa") (match_operand:VSX_F 2 "vsx_register_operand" ",0,wa,0") (match_operand:VSX_F 3 "vsx_register_operand" "0,,0,wa"))))] "VECTOR_UNIT_VSX_P (mode)" "@ xnmadda %x0,%x1,%x2 xnmaddm %x0,%x1,%x3 xnmadda %x0,%x1,%x2 xnmaddm %x0,%x1,%x3" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_nfmsdf4" [(set (match_operand:DF 0 "vsx_register_operand" "=ws,ws,?wa,?wa,d") (neg:DF (fma:DF (match_operand:DF 1 "vsx_register_operand" "%ws,ws,wa,wa,d") (match_operand:DF 2 "vsx_register_operand" "ws,0,wa,0,d") (neg:DF (match_operand:DF 3 "vsx_register_operand" "0,ws,0,wa,d")))))] "VECTOR_UNIT_VSX_P (DFmode)" "@ xsnmsubadp %x0,%x1,%x2 xsnmsubmdp %x0,%x1,%x3 xsnmsubadp %x0,%x1,%x2 xsnmsubmdp %x0,%x1,%x3 {fnms|fnmsub} %0,%1,%2,%3" [(set_attr "type" "fp") (set_attr "fp_type" "fp_maddsub_d")]) (define_insn "*vsx_nfmsv4sf4" [(set (match_operand:V4SF 0 "vsx_register_operand" "=wf,wf,?wa,?wa,v") (neg:V4SF (fma:V4SF (match_operand:V4SF 1 "vsx_register_operand" "%wf,wf,wa,wa,v") (match_operand:V4SF 2 "vsx_register_operand" "wf,0,wa,0,v") (neg:V4SF (match_operand:V4SF 3 "vsx_register_operand" "0,wf,0,wa,v")))))] "VECTOR_UNIT_VSX_P (V4SFmode)" "@ xvnmsubasp %x0,%x1,%x2 xvnmsubmsp %x0,%x1,%x3 xvnmsubasp %x0,%x1,%x2 xvnmsubmsp %x0,%x1,%x3 vnmsubfp %0,%1,%2,%3" [(set_attr "type" "vecfloat")]) (define_insn "*vsx_nfmsv2df4" [(set (match_operand:V2DF 0 "vsx_register_operand" "=wd,wd,?wa,?wa") (neg:V2DF (fma:V2DF (match_operand:V2DF 1 "vsx_register_operand" "%wd,wd,wa,wa") (match_operand:V2DF 2 "vsx_register_operand" "wd,0,wa,0") (neg:V2DF (match_operand:V2DF 3 "vsx_register_operand" "0,wd,0,wa")))))] "VECTOR_UNIT_VSX_P (V2DFmode)" "@ xvnmsubadp %x0,%x1,%x2 xvnmsubmdp %x0,%x1,%x3 xvnmsubadp %x0,%x1,%x2 xvnmsubmdp %x0,%x1,%x3" [(set_attr "type" "vecdouble")]) ;; Vector conditional expressions (no scalar version for these instructions) (define_insn "vsx_eq" [(set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (eq:VSX_F (match_operand:VSX_F 1 "vsx_register_operand" ",wa") (match_operand:VSX_F 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpeq %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_gt" [(set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (gt:VSX_F (match_operand:VSX_F 1 "vsx_register_operand" ",wa") (match_operand:VSX_F 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpgt %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_ge" [(set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (ge:VSX_F (match_operand:VSX_F 1 "vsx_register_operand" ",wa") (match_operand:VSX_F 2 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpge %x0,%x1,%x2" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; Floating point scalar compare (define_insn "*vsx_cmpdf_internal1" [(set (match_operand:CCFP 0 "cc_reg_operand" "=y,?y") (compare:CCFP (match_operand:DF 1 "gpc_reg_operand" "ws,wa") (match_operand:DF 2 "gpc_reg_operand" "ws,wa")))] "TARGET_HARD_FLOAT && TARGET_FPRS && TARGET_DOUBLE_FLOAT && VECTOR_UNIT_VSX_P (DFmode)" "xscmpudp %0,%x1,%x2" [(set_attr "type" "fpcompare")]) ;; Compare vectors producing a vector result and a predicate, setting CR6 to ;; indicate a combined status (define_insn "*vsx_eq__p" [(set (reg:CC 74) (unspec:CC [(eq:CC (match_operand:VSX_F 1 "vsx_register_operand" ",?wa") (match_operand:VSX_F 2 "vsx_register_operand" ",?wa"))] UNSPEC_PREDICATE)) (set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (eq:VSX_F (match_dup 1) (match_dup 2)))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpeq. %x0,%x1,%x2" [(set_attr "type" "")]) (define_insn "*vsx_gt__p" [(set (reg:CC 74) (unspec:CC [(gt:CC (match_operand:VSX_F 1 "vsx_register_operand" ",?wa") (match_operand:VSX_F 2 "vsx_register_operand" ",?wa"))] UNSPEC_PREDICATE)) (set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (gt:VSX_F (match_dup 1) (match_dup 2)))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpgt. %x0,%x1,%x2" [(set_attr "type" "")]) (define_insn "*vsx_ge__p" [(set (reg:CC 74) (unspec:CC [(ge:CC (match_operand:VSX_F 1 "vsx_register_operand" ",?wa") (match_operand:VSX_F 2 "vsx_register_operand" ",?wa"))] UNSPEC_PREDICATE)) (set (match_operand:VSX_F 0 "vsx_register_operand" "=,?wa") (ge:VSX_F (match_dup 1) (match_dup 2)))] "VECTOR_UNIT_VSX_P (mode)" "xvcmpge. %x0,%x1,%x2" [(set_attr "type" "")]) ;; Vector select (define_insn "*vsx_xxsel" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (if_then_else:VSX_L (ne:CC (match_operand:VSX_L 1 "vsx_register_operand" ",wa") (match_operand:VSX_L 4 "zero_constant" "")) (match_operand:VSX_L 2 "vsx_register_operand" ",wa") (match_operand:VSX_L 3 "vsx_register_operand" ",wa")))] "VECTOR_MEM_VSX_P (mode)" "xxsel %x0,%x3,%x2,%x1" [(set_attr "type" "vecperm")]) (define_insn "*vsx_xxsel_uns" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (if_then_else:VSX_L (ne:CCUNS (match_operand:VSX_L 1 "vsx_register_operand" ",wa") (match_operand:VSX_L 4 "zero_constant" "")) (match_operand:VSX_L 2 "vsx_register_operand" ",wa") (match_operand:VSX_L 3 "vsx_register_operand" ",wa")))] "VECTOR_MEM_VSX_P (mode)" "xxsel %x0,%x3,%x2,%x1" [(set_attr "type" "vecperm")]) ;; Copy sign (define_insn "vsx_copysign3" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa") (match_operand:VSX_B 2 "vsx_register_operand" ",wa")] UNSPEC_COPYSIGN))] "VECTOR_UNIT_VSX_P (mode)" "xcpsgn %x0,%x2,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; For the conversions, limit the register class for the integer value to be ;; the fprs because we don't want to add the altivec registers to movdi/movsi. ;; For the unsigned tests, there isn't a generic double -> unsigned conversion ;; in rs6000.md so don't test VECTOR_UNIT_VSX_P, just test against VSX. ;; Don't use vsx_register_operand here, use gpc_reg_operand to match rs6000.md. (define_insn "vsx_float2" [(set (match_operand:VSX_B 0 "gpc_reg_operand" "=,?wa") (float:VSX_B (match_operand: 1 "gpc_reg_operand" ",")))] "VECTOR_UNIT_VSX_P (mode)" "xcvsx %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_floatuns2" [(set (match_operand:VSX_B 0 "gpc_reg_operand" "=,?wa") (unsigned_float:VSX_B (match_operand: 1 "gpc_reg_operand" ",")))] "VECTOR_UNIT_VSX_P (mode)" "xcvux %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_fix_trunc2" [(set (match_operand: 0 "gpc_reg_operand" "=,?") (fix: (match_operand:VSX_B 1 "gpc_reg_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xcvsxs %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_fixuns_trunc2" [(set (match_operand: 0 "gpc_reg_operand" "=,?") (unsigned_fix: (match_operand:VSX_B 1 "gpc_reg_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xcvuxs %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; Math rounding functions (define_insn "vsx_xri" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_VSX_ROUND_I))] "VECTOR_UNIT_VSX_P (mode)" "xri %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_xric" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_VSX_ROUND_IC))] "VECTOR_UNIT_VSX_P (mode)" "xric %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_btrunc2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (fix:VSX_B (match_operand:VSX_B 1 "vsx_register_operand" ",wa")))] "VECTOR_UNIT_VSX_P (mode)" "xriz %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "*vsx_b2trunc2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_FRIZ))] "VECTOR_UNIT_VSX_P (mode)" "xriz %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_floor2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_FRIM))] "VECTOR_UNIT_VSX_P (mode)" "xrim %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) (define_insn "vsx_ceil2" [(set (match_operand:VSX_B 0 "vsx_register_operand" "=,?wa") (unspec:VSX_B [(match_operand:VSX_B 1 "vsx_register_operand" ",wa")] UNSPEC_FRIP))] "VECTOR_UNIT_VSX_P (mode)" "xrip %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; VSX convert to/from double vector ;; Convert between single and double precision ;; Don't use xscvspdp and xscvdpsp for scalar conversions, since the normal ;; scalar single precision instructions internally use the double format. ;; Prefer the altivec registers, since we likely will need to do a vperm (define_insn "vsx_" [(set (match_operand: 0 "vsx_register_operand" "=,?wa") (unspec: [(match_operand:VSX_SPDP 1 "vsx_register_operand" ",wa")] UNSPEC_VSX_CVSPDP))] "VECTOR_UNIT_VSX_P (mode)" " %x0,%x1" [(set_attr "type" "")]) ;; xscvspdp, represent the scalar SF type as V4SF (define_insn "vsx_xscvspdp" [(set (match_operand:DF 0 "vsx_register_operand" "=ws,?wa") (unspec:DF [(match_operand:V4SF 1 "vsx_register_operand" "wa,wa")] UNSPEC_VSX_CVSPDP))] "VECTOR_UNIT_VSX_P (V4SFmode)" "xscvspdp %x0,%x1" [(set_attr "type" "fp")]) ;; xscvdpsp used for splat'ing a scalar to V4SF, knowing that the internal SF ;; format of scalars is actually DF. (define_insn "vsx_xscvdpsp_scalar" [(set (match_operand:V4SF 0 "vsx_register_operand" "=wa") (unspec:V4SF [(match_operand:SF 1 "vsx_register_operand" "f")] UNSPEC_VSX_CVSPDP))] "VECTOR_UNIT_VSX_P (V4SFmode)" "xscvdpsp %x0,%x1" [(set_attr "type" "fp")]) ;; Same as vsx_xscvspdp, but use SF as the type (define_insn "vsx_xscvspdp_scalar2" [(set (match_operand:SF 0 "vsx_register_operand" "=f") (unspec:SF [(match_operand:V4SF 1 "vsx_register_operand" "wa")] UNSPEC_VSX_CVSPDP))] "VECTOR_UNIT_VSX_P (V4SFmode)" "xscvspdp %x0,%x1" [(set_attr "type" "fp")]) ;; Convert from 64-bit to 32-bit types ;; Note, favor the Altivec registers since the usual use of these instructions ;; is in vector converts and we need to use the Altivec vperm instruction. (define_insn "vsx_xvcvdpsxws" [(set (match_operand:V4SI 0 "vsx_register_operand" "=v,?wa") (unspec:V4SI [(match_operand:V2DF 1 "vsx_register_operand" "wd,wa")] UNSPEC_VSX_CVDPSXWS))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvdpsxws %x0,%x1" [(set_attr "type" "vecdouble")]) (define_insn "vsx_xvcvdpuxws" [(set (match_operand:V4SI 0 "vsx_register_operand" "=v,?wa") (unspec:V4SI [(match_operand:V2DF 1 "vsx_register_operand" "wd,wa")] UNSPEC_VSX_CVDPUXWS))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvdpuxws %x0,%x1" [(set_attr "type" "vecdouble")]) (define_insn "vsx_xvcvsxdsp" [(set (match_operand:V4SI 0 "vsx_register_operand" "=wd,?wa") (unspec:V4SI [(match_operand:V2DF 1 "vsx_register_operand" "wf,wa")] UNSPEC_VSX_CVSXDSP))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvsxdsp %x0,%x1" [(set_attr "type" "vecfloat")]) (define_insn "vsx_xvcvuxdsp" [(set (match_operand:V4SI 0 "vsx_register_operand" "=wd,?wa") (unspec:V4SI [(match_operand:V2DF 1 "vsx_register_operand" "wf,wa")] UNSPEC_VSX_CVUXDSP))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvuxwdp %x0,%x1" [(set_attr "type" "vecdouble")]) ;; Convert from 32-bit to 64-bit types (define_insn "vsx_xvcvsxwdp" [(set (match_operand:V2DF 0 "vsx_register_operand" "=wd,?wa") (unspec:V2DF [(match_operand:V4SI 1 "vsx_register_operand" "wf,wa")] UNSPEC_VSX_CVSXWDP))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvsxwdp %x0,%x1" [(set_attr "type" "vecdouble")]) (define_insn "vsx_xvcvuxwdp" [(set (match_operand:V2DF 0 "vsx_register_operand" "=wd,?wa") (unspec:V2DF [(match_operand:V4SI 1 "vsx_register_operand" "wf,wa")] UNSPEC_VSX_CVUXWDP))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvuxwdp %x0,%x1" [(set_attr "type" "vecdouble")]) (define_insn "vsx_xvcvspsxds" [(set (match_operand:V2DI 0 "vsx_register_operand" "=v,?wa") (unspec:V2DI [(match_operand:V4SF 1 "vsx_register_operand" "wd,wa")] UNSPEC_VSX_CVSPSXDS))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvspsxds %x0,%x1" [(set_attr "type" "vecdouble")]) (define_insn "vsx_xvcvspuxds" [(set (match_operand:V2DI 0 "vsx_register_operand" "=v,?wa") (unspec:V2DI [(match_operand:V4SF 1 "vsx_register_operand" "wd,wa")] UNSPEC_VSX_CVSPUXDS))] "VECTOR_UNIT_VSX_P (V2DFmode)" "xvcvspuxds %x0,%x1" [(set_attr "type" "vecdouble")]) ;; Only optimize (float (fix x)) -> frz if we are in fast-math mode, since ;; since the xsrdpiz instruction does not truncate the value if the floating ;; point value is < LONG_MIN or > LONG_MAX. (define_insn "*vsx_float_fix_2" [(set (match_operand:VSX_DF 0 "vsx_register_operand" "=,?wa") (float:VSX_DF (fix: (match_operand:VSX_DF 1 "vsx_register_operand" ",?wa"))))] "TARGET_HARD_FLOAT && TARGET_FPRS && TARGET_DOUBLE_FLOAT && VECTOR_UNIT_VSX_P (mode) && flag_unsafe_math_optimizations && !flag_trapping_math && TARGET_FRIZ" "xriz %x0,%x1" [(set_attr "type" "") (set_attr "fp_type" "")]) ;; Logical and permute operations (define_insn "*vsx_and3" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (and:VSX_L (match_operand:VSX_L 1 "vsx_register_operand" ",?wa") (match_operand:VSX_L 2 "vsx_register_operand" ",?wa")))] "VECTOR_MEM_VSX_P (mode)" "xxland %x0,%x1,%x2" [(set_attr "type" "vecsimple")]) (define_insn "*vsx_ior3" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (ior:VSX_L (match_operand:VSX_L 1 "vsx_register_operand" ",?wa") (match_operand:VSX_L 2 "vsx_register_operand" ",?wa")))] "VECTOR_MEM_VSX_P (mode)" "xxlor %x0,%x1,%x2" [(set_attr "type" "vecsimple")]) (define_insn "*vsx_xor3" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (xor:VSX_L (match_operand:VSX_L 1 "vsx_register_operand" ",?wa") (match_operand:VSX_L 2 "vsx_register_operand" ",?wa")))] "VECTOR_MEM_VSX_P (mode)" "xxlxor %x0,%x1,%x2" [(set_attr "type" "vecsimple")]) (define_insn "*vsx_one_cmpl2" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (not:VSX_L (match_operand:VSX_L 1 "vsx_register_operand" ",?wa")))] "VECTOR_MEM_VSX_P (mode)" "xxlnor %x0,%x1,%x1" [(set_attr "type" "vecsimple")]) (define_insn "*vsx_nor3" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (not:VSX_L (ior:VSX_L (match_operand:VSX_L 1 "vsx_register_operand" ",?wa") (match_operand:VSX_L 2 "vsx_register_operand" ",?wa"))))] "VECTOR_MEM_VSX_P (mode)" "xxlnor %x0,%x1,%x2" [(set_attr "type" "vecsimple")]) (define_insn "*vsx_andc3" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=,?wa") (and:VSX_L (not:VSX_L (match_operand:VSX_L 2 "vsx_register_operand" ",?wa")) (match_operand:VSX_L 1 "vsx_register_operand" ",?wa")))] "VECTOR_MEM_VSX_P (mode)" "xxlandc %x0,%x1,%x2" [(set_attr "type" "vecsimple")]) ;; Permute operations ;; Build a V2DF/V2DI vector from two scalars (define_insn "vsx_concat_" [(set (match_operand:VSX_D 0 "vsx_register_operand" "=wd,?wa") (unspec:VSX_D [(match_operand: 1 "vsx_register_operand" "ws,wa") (match_operand: 2 "vsx_register_operand" "ws,wa")] UNSPEC_VSX_CONCAT))] "VECTOR_MEM_VSX_P (mode)" "xxpermdi %x0,%x1,%x2,0" [(set_attr "type" "vecperm")]) ;; Special purpose concat using xxpermdi to glue two single precision values ;; together, relying on the fact that internally scalar floats are represented ;; as doubles. This is used to initialize a V4SF vector with 4 floats (define_insn "vsx_concat_v2sf" [(set (match_operand:V2DF 0 "vsx_register_operand" "=wd,?wa") (unspec:V2DF [(match_operand:SF 1 "vsx_register_operand" "f,f") (match_operand:SF 2 "vsx_register_operand" "f,f")] UNSPEC_VSX_CONCAT))] "VECTOR_MEM_VSX_P (V2DFmode)" "xxpermdi %x0,%x1,%x2,0" [(set_attr "type" "vecperm")]) ;; Set the element of a V2DI/VD2F mode (define_insn "vsx_set_" [(set (match_operand:VSX_D 0 "vsx_register_operand" "=wd,?wa") (unspec:VSX_D [(match_operand:VSX_D 1 "vsx_register_operand" "wd,wa") (match_operand: 2 "vsx_register_operand" "ws,wa") (match_operand:QI 3 "u5bit_cint_operand" "i,i")] UNSPEC_VSX_SET))] "VECTOR_MEM_VSX_P (mode)" { if (INTVAL (operands[3]) == 0) return \"xxpermdi %x0,%x2,%x1,1\"; else if (INTVAL (operands[3]) == 1) return \"xxpermdi %x0,%x1,%x2,0\"; else gcc_unreachable (); } [(set_attr "type" "vecperm")]) ;; Extract a DF/DI element from V2DF/V2DI (define_insn "vsx_extract_" [(set (match_operand: 0 "vsx_register_operand" "=ws,d,?wa") (vec_select: (match_operand:VSX_D 1 "vsx_register_operand" "wd,wd,wa") (parallel [(match_operand:QI 2 "u5bit_cint_operand" "i,i,i")])))] "VECTOR_MEM_VSX_P (mode)" { gcc_assert (UINTVAL (operands[2]) <= 1); operands[3] = GEN_INT (INTVAL (operands[2]) << 1); return \"xxpermdi %x0,%x1,%x1,%3\"; } [(set_attr "type" "vecperm")]) ;; Optimize extracting element 0 from memory (define_insn "*vsx_extract__zero" [(set (match_operand: 0 "vsx_register_operand" "=ws,d,?wa") (vec_select: (match_operand:VSX_D 1 "indexed_or_indirect_operand" "Z,Z,Z") (parallel [(const_int 0)])))] "VECTOR_MEM_VSX_P (mode) && WORDS_BIG_ENDIAN" "lxsd%U1x %x0,%y1" [(set_attr "type" "fpload") (set_attr "length" "4")]) ;; Extract a SF element from V4SF (define_insn_and_split "vsx_extract_v4sf" [(set (match_operand:SF 0 "vsx_register_operand" "=f,f") (vec_select:SF (match_operand:V4SF 1 "vsx_register_operand" "wa,wa") (parallel [(match_operand:QI 2 "u5bit_cint_operand" "O,i")]))) (clobber (match_scratch:V4SF 3 "=X,0"))] "VECTOR_UNIT_VSX_P (V4SFmode)" "@ xscvspdp %x0,%x1 #" "" [(const_int 0)] " { rtx op0 = operands[0]; rtx op1 = operands[1]; rtx op2 = operands[2]; rtx op3 = operands[3]; rtx tmp; HOST_WIDE_INT ele = INTVAL (op2); if (ele == 0) tmp = op1; else { if (GET_CODE (op3) == SCRATCH) op3 = gen_reg_rtx (V4SFmode); emit_insn (gen_vsx_xxsldwi_v4sf (op3, op1, op1, op2)); tmp = op3; } emit_insn (gen_vsx_xscvspdp_scalar2 (op0, tmp)); DONE; }" [(set_attr "length" "4,8") (set_attr "type" "fp")]) ;; Expand the builtin form of xxpermdi to canonical rtl. (define_expand "vsx_xxpermdi_" [(match_operand:VSX_L 0 "vsx_register_operand" "") (match_operand:VSX_L 1 "vsx_register_operand" "") (match_operand:VSX_L 2 "vsx_register_operand" "") (match_operand:QI 3 "u5bit_cint_operand" "")] "VECTOR_MEM_VSX_P (mode)" { rtx target = operands[0]; rtx op0 = operands[1]; rtx op1 = operands[2]; int mask = INTVAL (operands[3]); rtx perm0 = GEN_INT ((mask >> 1) & 1); rtx perm1 = GEN_INT ((mask & 1) + 2); rtx (*gen) (rtx, rtx, rtx, rtx, rtx); if (mode == V2DFmode) gen = gen_vsx_xxpermdi2_v2df_1; else { gen = gen_vsx_xxpermdi2_v2di_1; if (mode != V2DImode) { target = gen_lowpart (V2DImode, target); op0 = gen_lowpart (V2DImode, target); op1 = gen_lowpart (V2DImode, target); } } emit_insn (gen (target, op0, op1, perm0, perm1)); DONE; }) (define_insn "vsx_xxpermdi2__1" [(set (match_operand:VSX_D 0 "vsx_register_operand" "=wd") (vec_select:VSX_D (vec_concat: (match_operand:VSX_D 1 "vsx_register_operand" "wd") (match_operand:VSX_D 2 "vsx_register_operand" "wd")) (parallel [(match_operand 3 "const_0_to_1_operand" "") (match_operand 4 "const_2_to_3_operand" "")])))] "VECTOR_MEM_VSX_P (mode)" { int mask = (INTVAL (operands[3]) << 1) | (INTVAL (operands[4]) - 2); operands[3] = GEN_INT (mask); return "xxpermdi %x0,%x1,%x2,%3"; } [(set_attr "type" "vecperm")]) (define_expand "vec_perm_const" [(match_operand:VSX_D 0 "vsx_register_operand" "") (match_operand:VSX_D 1 "vsx_register_operand" "") (match_operand:VSX_D 2 "vsx_register_operand" "") (match_operand:V2DI 3 "" "")] "VECTOR_MEM_VSX_P (mode)" { if (rs6000_expand_vec_perm_const (operands)) DONE; else FAIL; }) ;; Expanders for builtins (define_expand "vsx_mergel_" [(set (match_operand:VSX_D 0 "vsx_register_operand" "") (vec_select:VSX_D (vec_concat: (match_operand:VSX_D 1 "vsx_register_operand" "") (match_operand:VSX_D 2 "vsx_register_operand" "")) (parallel [(const_int 1) (const_int 3)])))] "VECTOR_MEM_VSX_P (mode)" "") (define_expand "vsx_mergeh_" [(set (match_operand:VSX_D 0 "vsx_register_operand" "") (vec_select:VSX_D (vec_concat: (match_operand:VSX_D 1 "vsx_register_operand" "") (match_operand:VSX_D 2 "vsx_register_operand" "")) (parallel [(const_int 0) (const_int 2)])))] "VECTOR_MEM_VSX_P (mode)" "") ;; V2DF/V2DI splat (define_insn "vsx_splat_" [(set (match_operand:VSX_D 0 "vsx_register_operand" "=wd,wd,wd,?wa,?wa,?wa") (vec_duplicate:VSX_D (match_operand: 1 "splat_input_operand" "ws,f,Z,wa,wa,Z")))] "VECTOR_MEM_VSX_P (mode)" "@ xxpermdi %x0,%x1,%x1,0 xxpermdi %x0,%x1,%x1,0 lxvdsx %x0,%y1 xxpermdi %x0,%x1,%x1,0 xxpermdi %x0,%x1,%x1,0 lxvdsx %x0,%y1" [(set_attr "type" "vecperm,vecperm,vecload,vecperm,vecperm,vecload")]) ;; V4SF/V4SI splat (define_insn "vsx_xxspltw_" [(set (match_operand:VSX_W 0 "vsx_register_operand" "=wf,?wa") (vec_duplicate:VSX_W (vec_select: (match_operand:VSX_W 1 "vsx_register_operand" "wf,wa") (parallel [(match_operand:QI 2 "u5bit_cint_operand" "i,i")]))))] "VECTOR_MEM_VSX_P (mode)" "xxspltw %x0,%x1,%2" [(set_attr "type" "vecperm")]) ;; V4SF/V4SI interleave (define_insn "vsx_xxmrghw_" [(set (match_operand:VSX_W 0 "vsx_register_operand" "=wf,?wa") (vec_select:VSX_W (vec_concat: (match_operand:VSX_W 1 "vsx_register_operand" "wf,wa") (match_operand:VSX_W 2 "vsx_register_operand" "wf,wa")) (parallel [(const_int 0) (const_int 4) (const_int 1) (const_int 5)])))] "VECTOR_MEM_VSX_P (mode)" "xxmrghw %x0,%x1,%x2" [(set_attr "type" "vecperm")]) (define_insn "vsx_xxmrglw_" [(set (match_operand:VSX_W 0 "vsx_register_operand" "=wf,?wa") (vec_select:VSX_W (vec_concat: (match_operand:VSX_W 1 "vsx_register_operand" "wf,wa") (match_operand:VSX_W 2 "vsx_register_operand" "wf,?wa")) (parallel [(const_int 2) (const_int 6) (const_int 3) (const_int 7)])))] "VECTOR_MEM_VSX_P (mode)" "xxmrglw %x0,%x1,%x2" [(set_attr "type" "vecperm")]) ;; Shift left double by word immediate (define_insn "vsx_xxsldwi_" [(set (match_operand:VSX_L 0 "vsx_register_operand" "=wa") (unspec:VSX_L [(match_operand:VSX_L 1 "vsx_register_operand" "wa") (match_operand:VSX_L 2 "vsx_register_operand" "wa") (match_operand:QI 3 "u5bit_cint_operand" "i")] UNSPEC_VSX_SLDWI))] "VECTOR_MEM_VSX_P (mode)" "xxsldwi %x0,%x1,%x2,%3" [(set_attr "type" "vecperm")]) ;; Vector reduction insns and splitters (define_insn_and_split "*vsx_reduc__v2df" [(set (match_operand:V2DF 0 "vfloat_operand" "=&wd,&?wa,wd,?wa") (VEC_reduc:V2DF (vec_concat:V2DF (vec_select:DF (match_operand:V2DF 1 "vfloat_operand" "wd,wa,wd,wa") (parallel [(const_int 1)])) (vec_select:DF (match_dup 1) (parallel [(const_int 0)]))) (match_dup 1))) (clobber (match_scratch:V2DF 2 "=0,0,&wd,&wa"))] "VECTOR_UNIT_VSX_P (V2DFmode)" "#" "" [(const_int 0)] " { rtx tmp = (GET_CODE (operands[2]) == SCRATCH) ? gen_reg_rtx (V2DFmode) : operands[2]; emit_insn (gen_vsx_xxsldwi_v2df (tmp, operands[1], operands[1], const2_rtx)); emit_insn (gen_v2df3 (operands[0], tmp, operands[1])); DONE; }" [(set_attr "length" "8") (set_attr "type" "veccomplex")]) (define_insn_and_split "*vsx_reduc__v4sf" [(set (match_operand:V4SF 0 "vfloat_operand" "=wf,?wa") (VEC_reduc:V4SF (unspec:V4SF [(const_int 0)] UNSPEC_REDUC) (match_operand:V4SF 1 "vfloat_operand" "wf,wa"))) (clobber (match_scratch:V4SF 2 "=&wf,&wa")) (clobber (match_scratch:V4SF 3 "=&wf,&wa"))] "VECTOR_UNIT_VSX_P (V4SFmode)" "#" "" [(const_int 0)] " { rtx op0 = operands[0]; rtx op1 = operands[1]; rtx tmp2, tmp3, tmp4; if (can_create_pseudo_p ()) { tmp2 = gen_reg_rtx (V4SFmode); tmp3 = gen_reg_rtx (V4SFmode); tmp4 = gen_reg_rtx (V4SFmode); } else { tmp2 = operands[2]; tmp3 = operands[3]; tmp4 = tmp2; } emit_insn (gen_vsx_xxsldwi_v4sf (tmp2, op1, op1, const2_rtx)); emit_insn (gen_v4sf3 (tmp3, tmp2, op1)); emit_insn (gen_vsx_xxsldwi_v4sf (tmp4, tmp3, tmp3, GEN_INT (3))); emit_insn (gen_v4sf3 (op0, tmp4, tmp3)); DONE; }" [(set_attr "length" "16") (set_attr "type" "veccomplex")]) ;; Combiner patterns with the vector reduction patterns that knows we can get ;; to the top element of the V2DF array without doing an extract. (define_insn_and_split "*vsx_reduc__v2df_scalar" [(set (match_operand:DF 0 "vfloat_operand" "=&ws,&?wa,ws,?wa") (vec_select:DF (VEC_reduc:V2DF (vec_concat:V2DF (vec_select:DF (match_operand:V2DF 1 "vfloat_operand" "wd,wa,wd,wa") (parallel [(const_int 1)])) (vec_select:DF (match_dup 1) (parallel [(const_int 0)]))) (match_dup 1)) (parallel [(const_int 1)]))) (clobber (match_scratch:DF 2 "=0,0,&wd,&wa"))] "VECTOR_UNIT_VSX_P (V2DFmode)" "#" "" [(const_int 0)] " { rtx hi = gen_highpart (DFmode, operands[1]); rtx lo = (GET_CODE (operands[2]) == SCRATCH) ? gen_reg_rtx (DFmode) : operands[2]; emit_insn (gen_vsx_extract_v2df (lo, operands[1], const1_rtx)); emit_insn (gen_df3 (operands[0], hi, lo)); DONE; }" [(set_attr "length" "8") (set_attr "type" "veccomplex")]) (define_insn_and_split "*vsx_reduc__v4sf_scalar" [(set (match_operand:SF 0 "vfloat_operand" "=f,?f") (vec_select:SF (VEC_reduc:V4SF (unspec:V4SF [(const_int 0)] UNSPEC_REDUC) (match_operand:V4SF 1 "vfloat_operand" "wf,wa")) (parallel [(const_int 3)]))) (clobber (match_scratch:V4SF 2 "=&wf,&wa")) (clobber (match_scratch:V4SF 3 "=&wf,&wa")) (clobber (match_scratch:V4SF 4 "=0,0"))] "VECTOR_UNIT_VSX_P (V4SFmode)" "#" "" [(const_int 0)] " { rtx op0 = operands[0]; rtx op1 = operands[1]; rtx tmp2, tmp3, tmp4, tmp5; if (can_create_pseudo_p ()) { tmp2 = gen_reg_rtx (V4SFmode); tmp3 = gen_reg_rtx (V4SFmode); tmp4 = gen_reg_rtx (V4SFmode); tmp5 = gen_reg_rtx (V4SFmode); } else { tmp2 = operands[2]; tmp3 = operands[3]; tmp4 = tmp2; tmp5 = operands[4]; } emit_insn (gen_vsx_xxsldwi_v4sf (tmp2, op1, op1, const2_rtx)); emit_insn (gen_v4sf3 (tmp3, tmp2, op1)); emit_insn (gen_vsx_xxsldwi_v4sf (tmp4, tmp3, tmp3, GEN_INT (3))); emit_insn (gen_v4sf3 (tmp5, tmp4, tmp3)); emit_insn (gen_vsx_xscvspdp_scalar2 (op0, tmp5)); DONE; }" [(set_attr "length" "20") (set_attr "type" "veccomplex")])