;;- Instruction patterns for the System z vector facility ;; Copyright (C) 2015-2020 Free Software Foundation, Inc. ;; Contributed by Andreas Krebbel (Andreas.Krebbel@de.ibm.com) ;; This file is part of GCC. ;; GCC is free software; you can redistribute it and/or modify it under ;; the terms of the GNU General Public License as published by the Free ;; Software Foundation; either version 3, or (at your option) any later ;; version. ;; GCC is distributed in the hope that it will be useful, but WITHOUT ANY ;; WARRANTY; without even the implied warranty of MERCHANTABILITY or ;; FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License ;; for more details. ;; You should have received a copy of the GNU General Public License ;; along with GCC; see the file COPYING3. If not see ;; . ; All vector modes supported in a vector register (define_mode_iterator V [V1QI V2QI V4QI V8QI V16QI V1HI V2HI V4HI V8HI V1SI V2SI V4SI V1DI V2DI V1SF V2SF V4SF V1DF V2DF]) (define_mode_iterator VT [V1QI V2QI V4QI V8QI V16QI V1HI V2HI V4HI V8HI V1SI V2SI V4SI V1DI V2DI V1SF V2SF V4SF V1DF V2DF V1TF V1TI TI]) ; All modes directly supported by the hardware having full vector reg size ; V_HW2 is duplicate of V_HW for having two iterators expanding ; independently e.g. vcond (define_mode_iterator V_HW [V16QI V8HI V4SI V2DI (V1TI "TARGET_VXE") V2DF (V4SF "TARGET_VXE") (V1TF "TARGET_VXE")]) (define_mode_iterator V_HW2 [V16QI V8HI V4SI V2DI V2DF (V4SF "TARGET_VXE") (V1TF "TARGET_VXE")]) (define_mode_iterator V_HW_64 [V2DI V2DF]) (define_mode_iterator VT_HW_HSDT [V8HI V4SI V4SF V2DI V2DF V1TI V1TF TI TF]) (define_mode_iterator V_HW_HSD [V8HI V4SI (V4SF "TARGET_VXE") V2DI V2DF]) ; Including TI for instructions that support it (va, vn, ...) (define_mode_iterator VT_HW [V16QI V8HI V4SI V2DI V2DF V1TI TI (V4SF "TARGET_VXE") (V1TF "TARGET_VXE")]) ; All full size integer vector modes supported in a vector register + TImode (define_mode_iterator VIT_HW [V16QI V8HI V4SI V2DI V1TI TI]) (define_mode_iterator VI_HW [V16QI V8HI V4SI V2DI]) (define_mode_iterator VI_HW_QHS [V16QI V8HI V4SI]) (define_mode_iterator VI_HW_HSD [V8HI V4SI V2DI]) (define_mode_iterator VI_HW_HS [V8HI V4SI]) (define_mode_iterator VI_HW_QH [V16QI V8HI]) (define_mode_iterator VI_HW_4 [V4SI V4SF]) ; All integer vector modes supported in a vector register + TImode (define_mode_iterator VIT [V1QI V2QI V4QI V8QI V16QI V1HI V2HI V4HI V8HI V1SI V2SI V4SI V1DI V2DI V1TI TI]) (define_mode_iterator VI [V1QI V2QI V4QI V8QI V16QI V1HI V2HI V4HI V8HI V1SI V2SI V4SI V1DI V2DI]) (define_mode_iterator VI_QHS [V1QI V2QI V4QI V8QI V16QI V1HI V2HI V4HI V8HI V1SI V2SI V4SI]) (define_mode_iterator VFT [(V1SF "TARGET_VXE") (V2SF "TARGET_VXE") (V4SF "TARGET_VXE") V1DF V2DF (V1TF "TARGET_VXE")]) ; FP vector modes directly supported by the HW. This does not include ; vector modes using only part of a vector register and should be used ; for instructions which might trigger IEEE exceptions. (define_mode_iterator VF_HW [(V4SF "TARGET_VXE") V2DF (V1TF "TARGET_VXE")]) (define_mode_iterator V_8 [V1QI]) (define_mode_iterator V_16 [V2QI V1HI]) (define_mode_iterator V_32 [V4QI V2HI V1SI V1SF]) (define_mode_iterator V_64 [V8QI V4HI V2SI V2SF V1DI V1DF]) (define_mode_iterator V_128 [V16QI V8HI V4SI V4SF V2DI V2DF V1TI V1TF]) (define_mode_iterator V_128_NOSINGLE [V16QI V8HI V4SI V4SF V2DI V2DF]) ; 32 bit int<->fp vector conversion instructions are available since VXE2 (z15). (define_mode_iterator VX_VEC_CONV_BFP [V2DF (V4SF "TARGET_VXE2")]) (define_mode_iterator VX_VEC_CONV_INT [V2DI (V4SI "TARGET_VXE2")]) ; Empty string for all but TImode. This is used to hide the TImode ; expander name in case it is defined already. See addti3 for an ; example. (define_mode_attr ti* [(V1QI "") (V2QI "") (V4QI "") (V8QI "") (V16QI "") (V1HI "") (V2HI "") (V4HI "") (V8HI "") (V1SI "") (V2SI "") (V4SI "") (V1DI "") (V2DI "") (V1TI "") (TI "*") (V1SF "") (V2SF "") (V4SF "") (V1DF "") (V2DF "") (V1TF "") (TF "")]) ; The element type of the vector. (define_mode_attr non_vec[(V1QI "QI") (V2QI "QI") (V4QI "QI") (V8QI "QI") (V16QI "QI") (V1HI "HI") (V2HI "HI") (V4HI "HI") (V8HI "HI") (V1SI "SI") (V2SI "SI") (V4SI "SI") (V1DI "DI") (V2DI "DI") (V1TI "TI") (TI "TI") (V1SF "SF") (V2SF "SF") (V4SF "SF") (V1DF "DF") (V2DF "DF") (V1TF "TF") (TF "TF")]) ; Like above, but in lower case. (define_mode_attr non_vec_l[(V1QI "qi") (V2QI "qi") (V4QI "qi") (V8QI "qi") (V16QI "qi") (V1HI "hi") (V2HI "hi") (V4HI "hi") (V8HI "hi") (V1SI "si") (V2SI "si") (V4SI "si") (V1DI "di") (V2DI "di") (V1TI "ti") (TI "ti") (V1SF "sf") (V2SF "sf") (V4SF "sf") (V1DF "df") (V2DF "df") (V1TF "tf") (TF "tf")]) ; The instruction suffix for integer instructions and instructions ; which do not care about whether it is floating point or integer. (define_mode_attr bhfgq[(V1QI "b") (V2QI "b") (V4QI "b") (V8QI "b") (V16QI "b") (V1HI "h") (V2HI "h") (V4HI "h") (V8HI "h") (V1SI "f") (V2SI "f") (V4SI "f") (V1DI "g") (V2DI "g") (V1TI "q") (TI "q") (V1SF "f") (V2SF "f") (V4SF "f") (V1DF "g") (V2DF "g") (V1TF "q")]) ; This is for vmalhw. It gets an 'w' attached to avoid confusion with ; multiply and add logical high vmalh. (define_mode_attr w [(V1QI "") (V2QI "") (V4QI "") (V8QI "") (V16QI "") (V1HI "w") (V2HI "w") (V4HI "w") (V8HI "w") (V1SI "") (V2SI "") (V4SI "") (V1DI "") (V2DI "")]) ; Resulting mode of a vector comparison. For floating point modes an ; integer vector mode with the same element size is picked. (define_mode_attr tointvec [(V1QI "V1QI") (V2QI "V2QI") (V4QI "V4QI") (V8QI "V8QI") (V16QI "V16QI") (V1HI "V1HI") (V2HI "V2HI") (V4HI "V4HI") (V8HI "V8HI") (V1SI "V1SI") (V2SI "V2SI") (V4SI "V4SI") (V1DI "V1DI") (V2DI "V2DI") (V1TI "V1TI") (V1SF "V1SI") (V2SF "V2SI") (V4SF "V4SI") (V1DF "V1DI") (V2DF "V2DI") (V1TF "V1TI")]) (define_mode_attr vw [(SF "w") (V1SF "w") (V2SF "v") (V4SF "v") (DF "w") (V1DF "w") (V2DF "v") (TF "w") (V1TF "w")]) (define_mode_attr sdx [(SF "s") (V1SF "s") (V2SF "s") (V4SF "s") (DF "d") (V1DF "d") (V2DF "d") (TF "x") (V1TF "x")]) ; Vector with doubled element size. (define_mode_attr vec_double [(V1QI "V1HI") (V2QI "V1HI") (V4QI "V2HI") (V8QI "V4HI") (V16QI "V8HI") (V1HI "V1SI") (V2HI "V1SI") (V4HI "V2SI") (V8HI "V4SI") (V1SI "V1DI") (V2SI "V1DI") (V4SI "V2DI") (V1DI "V1TI") (V2DI "V1TI") (V1SF "V1DF") (V2SF "V1DF") (V4SF "V2DF")]) ; Vector with half the element size. (define_mode_attr vec_half [(V1HI "V2QI") (V2HI "V4QI") (V4HI "V8QI") (V8HI "V16QI") (V1SI "V2HI") (V2SI "V4HI") (V4SI "V8HI") (V1DI "V2SI") (V2DI "V4SI") (V1TI "V2DI") (V1DF "V2SF") (V2DF "V4SF") (V1TF "V1DF")]) ; Vector with half the element size AND half the number of elements. (define_mode_attr vec_halfhalf [(V2HI "V2QI") (V4HI "V4QI") (V8HI "V8QI") (V2SI "V2HI") (V4SI "V4HI") (V2DI "V2SI") (V2DF "V2SF")]) (define_mode_attr vec_halfnumelts [(V4SF "V2SF") (V4SI "V2SI")]) ; Comparison operators on int and fp compares which are directly ; supported by the HW. (define_code_iterator VICMP_HW_OP [eq gt gtu]) ; For int insn_cmp_op can be used in the insn name as well as in the asm output. (define_code_attr insn_cmp_op [(eq "eq") (gt "h") (gtu "hl") (ge "he")]) ; Flags for vector string instructions (vfae all 4, vfee only ZS and CS, vstrc all 4) (define_constants [(VSTRING_FLAG_IN 8) ; invert result (VSTRING_FLAG_RT 4) ; result type (VSTRING_FLAG_ZS 2) ; zero search (VSTRING_FLAG_CS 1)]) ; condition code set (include "vx-builtins.md") ; Full HW vector size moves ; We don't use lm/stm for 128 bit moves since these are slower than ; splitting it into separate moves. ; FIXME: More constants are possible by enabling jxx, jyy constraints ; for TImode (use double-int for the calculations) ; vgmb, vgmh, vgmf, vgmg, vrepib, vrepih, vrepif, vrepig (define_insn "mov" [(set (match_operand:V_128 0 "nonimmediate_operand" "=v,v,R, v, v, v, v, v,v,*d,*d,?o") (match_operand:V_128 1 "general_operand" " v,R,v,j00,jm1,jyy,jxx,jKK,d, v,dT,*d"))] "" "@ vlr\t%v0,%v1 vl\t%v0,%1%A1 vst\t%v1,%0%A0 vzero\t%v0 vone\t%v0 vgbm\t%v0,%t1 vgm\t%v0,%s1,%e1 vrepi\t%v0,%h1 vlvgp\t%v0,%1,%N1 # # #" [(set_attr "cpu_facility" "vx,vx,vx,vx,vx,vx,vx,vx,vx,vx,*,*") (set_attr "op_type" "VRR,VRX,VRX,VRI,VRI,VRI,VRI,VRI,VRR,*,*,*")]) ; VR -> GPR, no instruction so split it into 64 element sets. (define_split [(set (match_operand:V_128 0 "register_operand" "") (match_operand:V_128 1 "register_operand" ""))] "TARGET_VX && GENERAL_REG_P (operands[0]) && VECTOR_REG_P (operands[1])" [(set (match_dup 2) (unspec:DI [(subreg:V2DI (match_dup 1) 0) (const_int 0)] UNSPEC_VEC_EXTRACT)) (set (match_dup 3) (unspec:DI [(subreg:V2DI (match_dup 1) 0) (const_int 1)] UNSPEC_VEC_EXTRACT))] { operands[2] = operand_subword (operands[0], 0, 0, mode); operands[3] = operand_subword (operands[0], 1, 0, mode); }) ; Split the 128 bit GPR move into two word mode moves ; s390_split_ok_p decides which part needs to be moved first. (define_split [(set (match_operand:V_128 0 "nonimmediate_operand" "") (match_operand:V_128 1 "general_operand" ""))] "reload_completed && s390_split_ok_p (operands[0], operands[1], mode, 0)" [(set (match_dup 2) (match_dup 4)) (set (match_dup 3) (match_dup 5))] { operands[2] = operand_subword (operands[0], 0, 0, mode); operands[3] = operand_subword (operands[0], 1, 0, mode); operands[4] = operand_subword (operands[1], 0, 0, mode); operands[5] = operand_subword (operands[1], 1, 0, mode); }) (define_split [(set (match_operand:V_128 0 "nonimmediate_operand" "") (match_operand:V_128 1 "general_operand" ""))] "reload_completed && s390_split_ok_p (operands[0], operands[1], mode, 1)" [(set (match_dup 2) (match_dup 4)) (set (match_dup 3) (match_dup 5))] { operands[2] = operand_subword (operands[0], 1, 0, mode); operands[3] = operand_subword (operands[0], 0, 0, mode); operands[4] = operand_subword (operands[1], 1, 0, mode); operands[5] = operand_subword (operands[1], 0, 0, mode); }) ; This is the vector equivalent to the TImode splitter in s390.md. It ; is required if both target GPRs occur in the source address operand. ; For non-s_operands at least one of the target GPRs does not conflict ; with the address operand and one of the splitters above will take ; over. (define_split [(set (match_operand:V_128 0 "register_operand" "") (match_operand:V_128 1 "memory_operand" ""))] "TARGET_ZARCH && reload_completed && !VECTOR_REG_P (operands[0]) && !s_operand (operands[1], VOIDmode)" [(set (match_dup 0) (match_dup 1))] { rtx addr = operand_subword (operands[0], 1, 0, mode); addr = gen_lowpart (Pmode, addr); s390_load_address (addr, XEXP (operands[1], 0)); operands[1] = replace_equiv_address (operands[1], addr); }) ; Moves for smaller vector modes. ; In these patterns only the vlr, vone, and vzero instructions write ; VR bytes outside the mode. This should be ok since we disallow ; formerly bigger modes being accessed with smaller modes via ; subreg. Note: The vone, vzero instructions could easily be replaced ; with vlei which would only access the bytes belonging to the mode. ; However, this would probably be slower. (define_insn "mov" [(set (match_operand:V_8 0 "nonimmediate_operand" "=v,v,d,v,R, v, v, v, v,d, Q, S, Q, S, d, d,d,R,T") (match_operand:V_8 1 "general_operand" " v,d,v,R,v,j00,jm1,jyy,jxx,d,j00,j00,jm1,jm1,j00,jm1,T,d,d"))] "TARGET_VX" "@ vlr\t%v0,%v1 vlvgb\t%v0,%1,0 vlgvb\t%0,%v1,0 vleb\t%v0,%1,0 vsteb\t%v1,%0,0 vzero\t%v0 vone\t%v0 vgbm\t%v0,%t1 vgm\t%v0,%s1,%e1 lr\t%0,%1 mvi\t%0,0 mviy\t%0,0 mvi\t%0,-1 mviy\t%0,-1 lhi\t%0,0 lhi\t%0,-1 llc\t%0,%1 stc\t%1,%0 stcy\t%1,%0" [(set_attr "op_type" "VRR,VRS,VRS,VRX,VRX,VRI,VRI,VRI,VRI,RR,SI,SIY,SI,SIY,RI,RI,RXY,RX,RXY")]) (define_insn "mov" [(set (match_operand:V_16 0 "nonimmediate_operand" "=v,v,d,v,R, v, v, v, v,d, Q, Q, d, d,d,d,d,R,T,b") (match_operand:V_16 1 "general_operand" " v,d,v,R,v,j00,jm1,jyy,jxx,d,j00,jm1,j00,jm1,R,T,b,d,d,d"))] "" "@ vlr\t%v0,%v1 vlvgh\t%v0,%1,0 vlgvh\t%0,%v1,0 vleh\t%v0,%1,0 vsteh\t%v1,%0,0 vzero\t%v0 vone\t%v0 vgbm\t%v0,%t1 vgm\t%v0,%s1,%e1 lr\t%0,%1 mvhhi\t%0,0 mvhhi\t%0,-1 lhi\t%0,0 lhi\t%0,-1 lh\t%0,%1 lhy\t%0,%1 lhrl\t%0,%1 sth\t%1,%0 sthy\t%1,%0 sthrl\t%1,%0" [(set_attr "op_type" "VRR,VRS,VRS,VRX,VRX,VRI,VRI,VRI,VRI,RR,SIL,SIL,RI,RI,RX,RXY,RIL,RX,RXY,RIL")]) (define_insn "mov" [(set (match_operand:V_32 0 "nonimmediate_operand" "=f,f,f,R,T,v,v,d,v,R, f, v, v, v, v, Q, Q, d, d,d,d,d,d,R,T,b") (match_operand:V_32 1 "general_operand" " f,R,T,f,f,v,d,v,R,v,j00,j00,jm1,jyy,jxx,j00,jm1,j00,jm1,b,d,R,T,d,d,d"))] "TARGET_VX" "@ ldr\t%v0,%v1 lde\t%0,%1 ley\t%0,%1 ste\t%1,%0 stey\t%1,%0 vlr\t%v0,%v1 vlvgf\t%v0,%1,0 vlgvf\t%0,%v1,0 vlef\t%v0,%1,0 vstef\t%1,%0,0 lzer\t%v0 vzero\t%v0 vone\t%v0 vgbm\t%v0,%t1 vgm\t%v0,%s1,%e1 mvhi\t%0,0 mvhi\t%0,-1 lhi\t%0,0 lhi\t%0,-1 lrl\t%0,%1 lr\t%0,%1 l\t%0,%1 ly\t%0,%1 st\t%1,%0 sty\t%1,%0 strl\t%1,%0" [(set_attr "op_type" "RR,RXE,RXY,RX,RXY,VRR,VRS,VRS,VRX,VRX,RRE,VRI,VRI,VRI,VRI,SIL,SIL,RI,RI, RIL,RR,RX,RXY,RX,RXY,RIL")]) (define_insn "mov" [(set (match_operand:V_64 0 "nonimmediate_operand" "=f,f,f,R,T,v,v,d,v,R, f, v, v, v, v, Q, Q, d, d,f,d,d,d,d,T,b") (match_operand:V_64 1 "general_operand" " f,R,T,f,f,v,d,v,R,v,j00,j00,jm1,jyy,jxx,j00,jm1,j00,jm1,d,f,b,d,T,d,d"))] "TARGET_ZARCH" "@ ldr\t%0,%1 ld\t%0,%1 ldy\t%0,%1 std\t%1,%0 stdy\t%1,%0 vlr\t%v0,%v1 vlvgg\t%v0,%1,0 vlgvg\t%0,%v1,0 vleg\t%v0,%1,0 vsteg\t%v1,%0,0 lzdr\t%0 vzero\t%v0 vone\t%v0 vgbm\t%v0,%t1 vgm\t%v0,%s1,%e1 mvghi\t%0,0 mvghi\t%0,-1 lghi\t%0,0 lghi\t%0,-1 ldgr\t%0,%1 lgdr\t%0,%1 lgrl\t%0,%1 lgr\t%0,%1 lg\t%0,%1 stg\t%1,%0 stgrl\t%1,%0" [(set_attr "op_type" "RRE,RX,RXY,RX,RXY,VRR,VRS,VRS,VRX,VRX,RRE,VRI,VRI,VRI,VRI, SIL,SIL,RI,RI,RRE,RRE,RIL,RR,RXY,RXY,RIL")]) ; vec_load_lanes? ; vec_store_lanes? ; vec_set is supposed to *modify* an existing vector so operand 0 is ; duplicated as input operand. (define_expand "vec_set" [(set (match_operand:V 0 "register_operand" "") (unspec:V [(match_operand: 1 "general_operand" "") (match_operand:SI 2 "nonmemory_operand" "") (match_dup 0)] UNSPEC_VEC_SET))] "TARGET_VX") ; FIXME: Support also vector mode operands for 1 ; FIXME: A target memory operand seems to be useful otherwise we end ; up with vl vlvgg vst. Shouldn't the middle-end be able to handle ; that itself? ; vlvgb, vlvgh, vlvgf, vlvgg, vleb, vleh, vlef, vleg, vleib, vleih, vleif, vleig (define_insn "*vec_set" [(set (match_operand:V 0 "register_operand" "=v,v,v") (unspec:V [(match_operand: 1 "general_operand" "d,R,K") (match_operand:SI 2 "nonmemory_operand" "an,I,I") (match_operand:V 3 "register_operand" "0,0,0")] UNSPEC_VEC_SET))] "TARGET_VX && (!CONST_INT_P (operands[2]) || UINTVAL (operands[2]) < GET_MODE_NUNITS (mode))" "@ vlvg\t%v0,%1,%Y2 vle\t%v0,%1,%2 vlei\t%v0,%1,%2" [(set_attr "op_type" "VRS,VRX,VRI")]) ; vlvgb, vlvgh, vlvgf, vlvgg (define_insn "*vec_set_plus" [(set (match_operand:V 0 "register_operand" "=v") (unspec:V [(match_operand: 1 "general_operand" "d") (plus:SI (match_operand:SI 2 "register_operand" "a") (match_operand:SI 4 "const_int_operand" "n")) (match_operand:V 3 "register_operand" "0")] UNSPEC_VEC_SET))] "TARGET_VX" "vlvg\t%v0,%1,%Y4(%2)" [(set_attr "op_type" "VRS")]) ; FIXME: Support also vector mode operands for 0 ; FIXME: This should be (vec_select ..) or something but it does only allow constant selectors :( ; This is used via RTL standard name as well as for expanding the builtin (define_expand "vec_extract" [(set (match_operand: 0 "nonimmediate_operand" "") (unspec: [(match_operand:V 1 "register_operand" "") (match_operand:SI 2 "nonmemory_operand" "")] UNSPEC_VEC_EXTRACT))] "TARGET_VX") ; vlgvb, vlgvh, vlgvf, vlgvg, vsteb, vsteh, vstef, vsteg (define_insn "*vec_extract" [(set (match_operand: 0 "nonimmediate_operand" "=d,R") (unspec: [(match_operand:V 1 "register_operand" "v,v") (match_operand:SI 2 "nonmemory_operand" "an,I")] UNSPEC_VEC_EXTRACT))] "TARGET_VX && (!CONST_INT_P (operands[2]) || UINTVAL (operands[2]) < GET_MODE_NUNITS (mode))" "@ vlgv\t%0,%v1,%Y2 vste\t%v1,%0,%2" [(set_attr "op_type" "VRS,VRX")]) ; vlgvb, vlgvh, vlgvf, vlgvg (define_insn "*vec_extract_plus" [(set (match_operand: 0 "nonimmediate_operand" "=d") (unspec: [(match_operand:V 1 "register_operand" "v") (plus:SI (match_operand:SI 2 "nonmemory_operand" "a") (match_operand:SI 3 "const_int_operand" "n"))] UNSPEC_VEC_EXTRACT))] "TARGET_VX" "vlgv\t%0,%v1,%Y3(%2)" [(set_attr "op_type" "VRS")]) (define_expand "vec_init" [(match_operand:V_128 0 "register_operand" "") (match_operand:V_128 1 "nonmemory_operand" "")] "TARGET_VX" { s390_expand_vec_init (operands[0], operands[1]); DONE; }) (define_insn "*vec_vllezlf" [(set (match_operand:VI_HW_4 0 "register_operand" "=v") (vec_concat:VI_HW_4 (vec_concat: (match_operand: 1 "memory_operand" "R") (const_int 0)) (vec_concat: (const_int 0) (const_int 0))))] "TARGET_VXE" "vllezlf\t%v0,%1" [(set_attr "op_type" "VRX")]) ; Replicate from vector element ; vrepb, vreph, vrepf, vrepg (define_insn "*vec_splat" [(set (match_operand:V_128_NOSINGLE 0 "register_operand" "=v") (vec_duplicate:V_128_NOSINGLE (vec_select: (match_operand:V_128_NOSINGLE 1 "register_operand" "v") (parallel [(match_operand:QI 2 "const_mask_operand" "C")]))))] "TARGET_VX && UINTVAL (operands[2]) < GET_MODE_NUNITS (mode)" "vrep\t%v0,%v1,%2" [(set_attr "op_type" "VRI")]) ; vlrepb, vlreph, vlrepf, vlrepg, vrepib, vrepih, vrepif, vrepig, vrepb, vreph, vrepf, vrepg (define_insn "*vec_splats" [(set (match_operand:V_128_NOSINGLE 0 "register_operand" "=v,v,v,v") (vec_duplicate:V_128_NOSINGLE (match_operand: 1 "general_operand" " R,K,v,d")))] "TARGET_VX" "@ vlrep\t%v0,%1 vrepi\t%v0,%h1 vrep\t%v0,%v1,0 #" [(set_attr "op_type" "VRX,VRI,VRI,*")]) ; vlbrreph, vlbrrepf, vlbrrepg (define_insn "*vec_splats_bswap_vec" [(set (match_operand:V_HW_HSD 0 "register_operand" "=v") (bswap:V_HW_HSD (vec_duplicate:V_HW_HSD (match_operand: 1 "memory_operand" "R")))) (use (match_operand:V16QI 2 "permute_pattern_operand" "X"))] "TARGET_VXE2" "vlbrrep\t%v0,%1" [(set_attr "op_type" "VRX")]) ; Why do we need both? Shouldn't there be a canonical form? ; vlbrreph, vlbrrepf, vlbrrepg (define_insn "*vec_splats_bswap_elem" [(set (match_operand:V_HW_HSD 0 "register_operand" "=v") (vec_duplicate:V_HW_HSD (bswap: (match_operand: 1 "memory_operand" "R"))))] "TARGET_VXE2" "vlbrrep\t%v0,%1" [(set_attr "op_type" "VRX")]) ; A TFmode operand resides in FPR register pairs while V1TF is in a ; single vector register. (define_insn "*vec_tf_to_v1tf" [(set (match_operand:V1TF 0 "nonimmediate_operand" "=v,v,R,v,v") (vec_duplicate:V1TF (match_operand:TF 1 "general_operand" "v,R,v,G,d")))] "TARGET_VX" "@ vmrhg\t%v0,%1,%N1 vl\t%v0,%1%A1 vst\t%v1,%0%A0 vzero\t%v0 vlvgp\t%v0,%1,%N1" [(set_attr "op_type" "VRR,VRX,VRX,VRI,VRR")]) (define_insn "*vec_ti_to_v1ti" [(set (match_operand:V1TI 0 "nonimmediate_operand" "=v,v,R, v, v,v") (vec_duplicate:V1TI (match_operand:TI 1 "general_operand" "v,R,v,j00,jm1,d")))] "TARGET_VX" "@ vlr\t%v0,%v1 vl\t%v0,%1%A1 vst\t%v1,%0%A0 vzero\t%v0 vone\t%v0 vlvgp\t%v0,%1,%N1" [(set_attr "op_type" "VRR,VRX,VRX,VRI,VRI,VRR")]) ; vec_splats is supposed to replicate op1 into all elements of op0 ; This splitter first sets the rightmost element of op0 to op1 and ; then does a vec_splat to replicate that element into all other ; elements. (define_split [(set (match_operand:V_128_NOSINGLE 0 "register_operand" "") (vec_duplicate:V_128_NOSINGLE (match_operand: 1 "register_operand" "")))] "TARGET_VX && GENERAL_REG_P (operands[1])" [(set (match_dup 0) (unspec:V_128_NOSINGLE [(match_dup 1) (match_dup 2) (match_dup 0)] UNSPEC_VEC_SET)) (set (match_dup 0) (vec_duplicate:V_128_NOSINGLE (vec_select: (match_dup 0) (parallel [(match_dup 2)]))))] { operands[2] = GEN_INT (GET_MODE_NUNITS (mode) - 1); }) (define_predicate "vcond_comparison_operator" (match_operand 0 "comparison_operator") { if (!HONOR_NANS (GET_MODE (XEXP (op, 0))) && !HONOR_NANS (GET_MODE (XEXP (op, 1)))) return true; switch (GET_CODE (op)) { case LE: case LT: case GE: case GT: case LTGT: /* Signaling vector comparisons are supported only on z14+. */ return TARGET_VXE || TARGET_NONSIGNALING_VECTOR_COMPARE_OK; default: return true; } }) (define_expand "vcond" [(set (match_operand:V_HW 0 "register_operand" "") (if_then_else:V_HW (match_operator 3 "vcond_comparison_operator" [(match_operand:V_HW2 4 "register_operand" "") (match_operand:V_HW2 5 "nonmemory_operand" "")]) (match_operand:V_HW 1 "nonmemory_operand" "") (match_operand:V_HW 2 "nonmemory_operand" "")))] "TARGET_VX && GET_MODE_NUNITS (mode) == GET_MODE_NUNITS (mode)" { s390_expand_vcond (operands[0], operands[1], operands[2], GET_CODE (operands[3]), operands[4], operands[5]); DONE; }) (define_expand "vcondu" [(set (match_operand:V_HW 0 "register_operand" "") (if_then_else:V_HW (match_operator 3 "comparison_operator" [(match_operand:V_HW2 4 "register_operand" "") (match_operand:V_HW2 5 "nonmemory_operand" "")]) (match_operand:V_HW 1 "nonmemory_operand" "") (match_operand:V_HW 2 "nonmemory_operand" "")))] "TARGET_VX && GET_MODE_NUNITS (mode) == GET_MODE_NUNITS (mode)" { s390_expand_vcond (operands[0], operands[1], operands[2], GET_CODE (operands[3]), operands[4], operands[5]); DONE; }) ; We only have HW support for byte vectors. The middle-end is ; supposed to lower the mode if required. (define_insn "vec_permv16qi" [(set (match_operand:V16QI 0 "register_operand" "=v") (unspec:V16QI [(match_operand:V16QI 1 "register_operand" "v") (match_operand:V16QI 2 "register_operand" "v") (match_operand:V16QI 3 "register_operand" "v")] UNSPEC_VEC_PERM))] "TARGET_VX" "vperm\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) (define_insn "*vec_perm" [(set (match_operand:VT_HW 0 "register_operand" "=v") (subreg:VT_HW (unspec:V16QI [(subreg:V16QI (match_operand:VT_HW 1 "register_operand" "v") 0) (subreg:V16QI (match_operand:VT_HW 2 "register_operand" "v") 0) (match_operand:V16QI 3 "register_operand" "v")] UNSPEC_VEC_PERM) 0))] "TARGET_VX" "vperm\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) ; vec_perm_const for V2DI using vpdi? ;; ;; Vector integer arithmetic instructions ;; ; vab, vah, vaf, vag, vaq ; We use nonimmediate_operand instead of register_operand since it is ; better to have the reloads into VRs instead of splitting the ; operation into two DImode ADDs. (define_insn "add3" [(set (match_operand:VIT 0 "nonimmediate_operand" "=v") (plus:VIT (match_operand:VIT 1 "nonimmediate_operand" "v") (match_operand:VIT 2 "general_operand" "v")))] "TARGET_VX" "va\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vsb, vsh, vsf, vsg, vsq (define_insn "sub3" [(set (match_operand:VIT 0 "nonimmediate_operand" "=v") (minus:VIT (match_operand:VIT 1 "nonimmediate_operand" "v") (match_operand:VIT 2 "general_operand" "v")))] "TARGET_VX" "vs\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmlb, vmlhw, vmlf (define_insn "mul3" [(set (match_operand:VI_QHS 0 "register_operand" "=v") (mult:VI_QHS (match_operand:VI_QHS 1 "register_operand" "v") (match_operand:VI_QHS 2 "register_operand" "v")))] "TARGET_VX" "vml\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vlcb, vlch, vlcf, vlcg (define_insn "neg2" [(set (match_operand:VI 0 "register_operand" "=v") (neg:VI (match_operand:VI 1 "register_operand" "v")))] "TARGET_VX" "vlc\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; vlpb, vlph, vlpf, vlpg (define_insn "abs2" [(set (match_operand:VI 0 "register_operand" "=v") (abs:VI (match_operand:VI 1 "register_operand" "v")))] "TARGET_VX" "vlp\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; Vector sum across ; Sum across DImode parts of the 1st operand and add the rightmost ; element of 2nd operand ; vsumgh, vsumgf (define_insn "*vec_sum2" [(set (match_operand:V2DI 0 "register_operand" "=v") (unspec:V2DI [(match_operand:VI_HW_HS 1 "register_operand" "v") (match_operand:VI_HW_HS 2 "register_operand" "v")] UNSPEC_VEC_VSUMG))] "TARGET_VX" "vsumg\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vsumb, vsumh (define_insn "*vec_sum4" [(set (match_operand:V4SI 0 "register_operand" "=v") (unspec:V4SI [(match_operand:VI_HW_QH 1 "register_operand" "v") (match_operand:VI_HW_QH 2 "register_operand" "v")] UNSPEC_VEC_VSUM))] "TARGET_VX" "vsum\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ;; ;; Vector bit instructions (int + fp) ;; ; Vector and (define_insn "and3" [(set (match_operand:VT 0 "register_operand" "=v") (and:VT (match_operand:VT 1 "register_operand" "v") (match_operand:VT 2 "register_operand" "v")))] "TARGET_VX" "vn\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector not and (define_insn "notand3" [(set (match_operand:VT 0 "register_operand" "=v") (ior:VT (not:VT (match_operand:VT 1 "register_operand" "v")) (not:VT (match_operand:VT 2 "register_operand" "v"))))] "TARGET_VXE" "vnn\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector or (define_insn "ior3" [(set (match_operand:VT 0 "register_operand" "=v") (ior:VT (match_operand:VT 1 "register_operand" "v") (match_operand:VT 2 "register_operand" "v")))] "TARGET_VX" "vo\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector or with complement (define_insn "ior_not3" [(set (match_operand:VT 0 "register_operand" "=v") (ior:VT (not:VT (match_operand:VT 2 "register_operand" "v")) (match_operand:VT 1 "register_operand" "v")))] "TARGET_VXE" "voc\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector xor (define_insn "xor3" [(set (match_operand:VT 0 "register_operand" "=v") (xor:VT (match_operand:VT 1 "register_operand" "v") (match_operand:VT 2 "register_operand" "v")))] "TARGET_VX" "vx\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector not xor (define_insn "notxor3" [(set (match_operand:VT 0 "register_operand" "=v") (not:VT (xor:VT (match_operand:VT 1 "register_operand" "v") (match_operand:VT 2 "register_operand" "v"))))] "TARGET_VXE" "vnx\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Bitwise inversion of a vector (define_insn "one_cmpl2" [(set (match_operand:VT 0 "register_operand" "=v") (not:VT (match_operand:VT 1 "register_operand" "v")))] "TARGET_VX" "vnot\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; Vector population count (define_expand "popcount2" [(set (match_operand:VI_HW 0 "register_operand" "=v") (unspec:VI_HW [(match_operand:VI_HW 1 "register_operand" "v")] UNSPEC_POPCNT))] "TARGET_VX" { if (TARGET_VXE) emit_insn (gen_popcount2_vxe (operands[0], operands[1])); else emit_insn (gen_popcount2_vx (operands[0], operands[1])); DONE; }) ; vpopctb, vpopcth, vpopctf, vpopctg (define_insn "popcount2_vxe" [(set (match_operand:VI_HW 0 "register_operand" "=v") (unspec:VI_HW [(match_operand:VI_HW 1 "register_operand" "v")] UNSPEC_POPCNT))] "TARGET_VXE" "vpopct\t%v0,%v1" [(set_attr "op_type" "VRR")]) (define_insn "popcountv16qi2_vx" [(set (match_operand:V16QI 0 "register_operand" "=v") (unspec:V16QI [(match_operand:V16QI 1 "register_operand" "v")] UNSPEC_POPCNT))] "TARGET_VX && !TARGET_VXE" "vpopct\t%v0,%v1,0" [(set_attr "op_type" "VRR")]) ; vpopct only counts bits in byte elements. Bigger element sizes need ; to be emulated. Word and doubleword elements can use the sum across ; instructions. For halfword sized elements we do a shift of a copy ; of the result, add it to the result and extend it to halfword ; element size (unpack). (define_expand "popcountv8hi2_vx" [(set (match_dup 2) (unspec:V16QI [(match_operand:V8HI 1 "register_operand" "v")] UNSPEC_POPCNT)) ; Make a copy of the result (set (match_dup 3) (match_dup 2)) ; Generate the shift count operand in a VR (8->byte 7) (set (match_dup 4) (match_dup 5)) (set (match_dup 4) (unspec:V16QI [(const_int 8) (const_int 7) (match_dup 4)] UNSPEC_VEC_SET)) ; Vector shift right logical by one byte (set (match_dup 3) (unspec:V16QI [(match_dup 3) (match_dup 4)] UNSPEC_VEC_SRLB)) ; Add the shifted and the original result (set (match_dup 2) (plus:V16QI (match_dup 2) (match_dup 3))) ; Generate mask for the odd numbered byte elements (set (match_dup 3) (const_vector:V16QI [(const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255) (const_int 0) (const_int 255)])) ; Zero out the even indexed bytes (set (match_operand:V8HI 0 "register_operand" "=v") (and:V8HI (subreg:V8HI (match_dup 2) 0) (subreg:V8HI (match_dup 3) 0))) ] "TARGET_VX && !TARGET_VXE" { operands[1] = simplify_gen_subreg (V16QImode, operands[1], V8HImode, 0); operands[2] = gen_reg_rtx (V16QImode); operands[3] = gen_reg_rtx (V16QImode); operands[4] = gen_reg_rtx (V16QImode); operands[5] = CONST0_RTX (V16QImode); }) (define_expand "popcountv4si2_vx" [(set (match_dup 2) (unspec:V16QI [(match_operand:V4SI 1 "register_operand" "v")] UNSPEC_POPCNT)) (set (match_operand:V4SI 0 "register_operand" "=v") (unspec:V4SI [(match_dup 2) (match_dup 3)] UNSPEC_VEC_VSUM))] "TARGET_VX && !TARGET_VXE" { operands[1] = simplify_gen_subreg (V16QImode, operands[1], V4SImode, 0); operands[2] = gen_reg_rtx (V16QImode); operands[3] = force_reg (V16QImode, CONST0_RTX (V16QImode)); }) (define_expand "popcountv2di2_vx" [(set (match_dup 2) (unspec:V16QI [(match_operand:V2DI 1 "register_operand" "v")] UNSPEC_POPCNT)) (set (match_dup 3) (unspec:V4SI [(match_dup 2) (match_dup 4)] UNSPEC_VEC_VSUM)) (set (match_operand:V2DI 0 "register_operand" "=v") (unspec:V2DI [(match_dup 3) (match_dup 5)] UNSPEC_VEC_VSUMG))] "TARGET_VX && !TARGET_VXE" { operands[1] = simplify_gen_subreg (V16QImode, operands[1], V2DImode, 0); operands[2] = gen_reg_rtx (V16QImode); operands[3] = gen_reg_rtx (V4SImode); operands[4] = force_reg (V16QImode, CONST0_RTX (V16QImode)); operands[5] = force_reg (V4SImode, CONST0_RTX (V4SImode)); }) ; Count leading zeros ; vclzb, vclzh, vclzf, vclzg (define_insn "clz2" [(set (match_operand:V 0 "register_operand" "=v") (clz:V (match_operand:V 1 "register_operand" "v")))] "TARGET_VX" "vclz\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; Count trailing zeros ; vctzb, vctzh, vctzf, vctzg (define_insn "ctz2" [(set (match_operand:V 0 "register_operand" "=v") (ctz:V (match_operand:V 1 "register_operand" "v")))] "TARGET_VX" "vctz\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; Each vector element rotated by the corresponding vector element ; verllvb, verllvh, verllvf, verllvg (define_insn "vrotl3" [(set (match_operand:VI 0 "register_operand" "=v") (rotate:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "verllv\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector rotate and shift by scalar instructions (define_code_iterator VEC_SHIFTS [ashift ashiftrt lshiftrt rotate]) (define_code_attr vec_shifts_name [(ashift "ashl") (ashiftrt "ashr") (lshiftrt "lshr") (rotate "rotl")]) (define_code_attr vec_shifts_mnem [(ashift "vesl") (ashiftrt "vesra") (lshiftrt "vesrl") (rotate "verll")]) ; Each vector element rotated by a scalar (define_expand "3" [(set (match_operand:VI 0 "register_operand" "") (VEC_SHIFTS:VI (match_operand:VI 1 "register_operand" "") (match_operand:QI 2 "shift_count_operand" "")))] "TARGET_VX") ; verllb, verllh, verllf, verllg ; veslb, veslh, veslf, veslg ; vesrab, vesrah, vesraf, vesrag ; vesrlb, vesrlh, vesrlf, vesrlg (define_insn "*3" [(set (match_operand:VI 0 "register_operand" "=v") (VEC_SHIFTS:VI (match_operand:VI 1 "register_operand" "v") (match_operand:QI 2 "shift_count_operand_vec" "jsc")))] "TARGET_VX && s390_valid_shift_count (operands[2], GET_MODE_BITSIZE (GET_MODE_INNER (mode)) - 1) " "\t%v0,%v1,%Y2" [(set_attr "op_type" "VRS")]) ; Shift each element by corresponding vector element ; veslvb, veslvh, veslvf, veslvg (define_insn "vashl3" [(set (match_operand:VI 0 "register_operand" "=v") (ashift:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "veslv\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vesravb, vesravh, vesravf, vesravg (define_insn "vashr3" [(set (match_operand:VI 0 "register_operand" "=v") (ashiftrt:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vesrav\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vesrlvb, vesrlvh, vesrlvf, vesrlvg (define_insn "vlshr3" [(set (match_operand:VI 0 "register_operand" "=v") (lshiftrt:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vesrlv\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector shift right logical by byte ; Pattern used by e.g. popcount (define_insn "*vec_srb" [(set (match_operand:V_128 0 "register_operand" "=v") (unspec:V_128 [(match_operand:V_128 1 "register_operand" "v") (match_operand:V16QI 2 "register_operand" "v")] UNSPEC_VEC_SRLB))] "TARGET_VX" "vsrlb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Vector shift left by byte (define_insn "*vec_slb" [(set (match_operand:V_128 0 "register_operand" "=v") (unspec:V_128 [(match_operand:V_128 1 "register_operand" "v") (match_operand:V16QI 2 "register_operand" "v")] UNSPEC_VEC_SLB))] "TARGET_VX" "vslb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vec_shr is defined as shift towards element 0 ; this means it is a left shift on BE targets! (define_expand "vec_shr_" [(set (match_dup 3) (unspec:V16QI [(match_operand:SI 2 "const_shift_by_byte_operand" "") (const_int 7) (match_dup 3)] UNSPEC_VEC_SET)) (set (match_operand:V_128 0 "register_operand" "") (unspec:V_128 [(match_operand:V_128 1 "register_operand" "") (match_dup 3)] UNSPEC_VEC_SLB))] "TARGET_VX" { operands[3] = gen_reg_rtx(V16QImode); }) ; vmnb, vmnh, vmnf, vmng (define_insn "smin3" [(set (match_operand:VI 0 "register_operand" "=v") (smin:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vmn\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmxb, vmxh, vmxf, vmxg (define_insn "smax3" [(set (match_operand:VI 0 "register_operand" "=v") (smax:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vmx\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmnlb, vmnlh, vmnlf, vmnlg (define_insn "umin3" [(set (match_operand:VI 0 "register_operand" "=v") (umin:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vmnl\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmxlb, vmxlh, vmxlf, vmxlg (define_insn "umax3" [(set (match_operand:VI 0 "register_operand" "=v") (umax:VI (match_operand:VI 1 "register_operand" "v") (match_operand:VI 2 "register_operand" "v")))] "TARGET_VX" "vmxl\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmeb, vmeh, vmef (define_insn "vec_widen_smult_even_" [(set (match_operand: 0 "register_operand" "=v") (unspec: [(match_operand:VI_QHS 1 "register_operand" "v") (match_operand:VI_QHS 2 "register_operand" "v")] UNSPEC_VEC_SMULT_EVEN))] "TARGET_VX" "vme\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmleb, vmleh, vmlef (define_insn "vec_widen_umult_even_" [(set (match_operand: 0 "register_operand" "=v") (unspec: [(match_operand:VI_QHS 1 "register_operand" "v") (match_operand:VI_QHS 2 "register_operand" "v")] UNSPEC_VEC_UMULT_EVEN))] "TARGET_VX" "vmle\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmob, vmoh, vmof (define_insn "vec_widen_smult_odd_" [(set (match_operand: 0 "register_operand" "=v") (unspec: [(match_operand:VI_QHS 1 "register_operand" "v") (match_operand:VI_QHS 2 "register_operand" "v")] UNSPEC_VEC_SMULT_ODD))] "TARGET_VX" "vmo\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vmlob, vmloh, vmlof (define_insn "vec_widen_umult_odd_" [(set (match_operand: 0 "register_operand" "=v") (unspec: [(match_operand:VI_QHS 1 "register_operand" "v") (match_operand:VI_QHS 2 "register_operand" "v")] UNSPEC_VEC_UMULT_ODD))] "TARGET_VX" "vmlo\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Widening hi/lo multiplications ; The S/390 instructions vml and vmh return the low or high parts of ; the double sized result elements in the corresponding elements of ; the target register. That's NOT what the vec_widen_umult_lo/hi ; patterns are expected to do. ; We emulate the widening lo/hi multiplies with the even/odd versions ; followed by a vector merge (define_expand "vec_widen_umult_lo_" [(set (match_dup 3) (unspec: [(match_operand:VI_QHS 1 "register_operand" "") (match_operand:VI_QHS 2 "register_operand" "")] UNSPEC_VEC_UMULT_EVEN)) (set (match_dup 4) (unspec: [(match_dup 1) (match_dup 2)] UNSPEC_VEC_UMULT_ODD)) (set (match_operand: 0 "register_operand" "") (unspec: [(match_dup 3) (match_dup 4)] UNSPEC_VEC_MERGEL))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); operands[4] = gen_reg_rtx (mode); }) (define_expand "vec_widen_umult_hi_" [(set (match_dup 3) (unspec: [(match_operand:VI_QHS 1 "register_operand" "") (match_operand:VI_QHS 2 "register_operand" "")] UNSPEC_VEC_UMULT_EVEN)) (set (match_dup 4) (unspec: [(match_dup 1) (match_dup 2)] UNSPEC_VEC_UMULT_ODD)) (set (match_operand: 0 "register_operand" "") (unspec: [(match_dup 3) (match_dup 4)] UNSPEC_VEC_MERGEH))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); operands[4] = gen_reg_rtx (mode); }) (define_expand "vec_widen_smult_lo_" [(set (match_dup 3) (unspec: [(match_operand:VI_QHS 1 "register_operand" "") (match_operand:VI_QHS 2 "register_operand" "")] UNSPEC_VEC_SMULT_EVEN)) (set (match_dup 4) (unspec: [(match_dup 1) (match_dup 2)] UNSPEC_VEC_SMULT_ODD)) (set (match_operand: 0 "register_operand" "") (unspec: [(match_dup 3) (match_dup 4)] UNSPEC_VEC_MERGEL))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); operands[4] = gen_reg_rtx (mode); }) (define_expand "vec_widen_smult_hi_" [(set (match_dup 3) (unspec: [(match_operand:VI_QHS 1 "register_operand" "") (match_operand:VI_QHS 2 "register_operand" "")] UNSPEC_VEC_SMULT_EVEN)) (set (match_dup 4) (unspec: [(match_dup 1) (match_dup 2)] UNSPEC_VEC_SMULT_ODD)) (set (match_operand: 0 "register_operand" "") (unspec: [(match_dup 3) (match_dup 4)] UNSPEC_VEC_MERGEH))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); operands[4] = gen_reg_rtx (mode); }) ; vec_widen_ushiftl_hi ; vec_widen_ushiftl_lo ; vec_widen_sshiftl_hi ; vec_widen_sshiftl_lo ;; ;; Vector floating point arithmetic instructions ;; ; vfasb, vfadb, wfasb, wfadb, wfaxb (define_insn "add3" [(set (match_operand:VF_HW 0 "register_operand" "=v") (plus:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VX" "fab\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfssb, vfsdb, wfssb, wfsdb, wfsxb (define_insn "sub3" [(set (match_operand:VF_HW 0 "register_operand" "=v") (minus:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VX" "fsb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfmsb, vfmdb, wfmsb, wfmdb, wfmxb (define_insn "mul3" [(set (match_operand:VF_HW 0 "register_operand" "=v") (mult:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VX" "fmb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfdsb, vfddb, wfdsb, wfddb, wfdxb (define_insn "div3" [(set (match_operand:VF_HW 0 "register_operand" "=v") (div:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VX" "fdb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfsqsb, vfsqdb, wfsqsb, wfsqdb, wfsqxb (define_insn "sqrt2" [(set (match_operand:VF_HW 0 "register_operand" "=v") (sqrt:VF_HW (match_operand:VF_HW 1 "register_operand" "v")))] "TARGET_VX" "fsqb\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; vfmasb, vfmadb, wfmasb, wfmadb, wfmaxb (define_insn "fma4" [(set (match_operand:VF_HW 0 "register_operand" "=v") (fma:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v") (match_operand:VF_HW 3 "register_operand" "v")))] "TARGET_VX" "fmab\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) ; vfmssb, vfmsdb, wfmssb, wfmsdb, wfmsxb (define_insn "fms4" [(set (match_operand:VF_HW 0 "register_operand" "=v") (fma:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v") (neg:VF_HW (match_operand:VF_HW 3 "register_operand" "v"))))] "TARGET_VX" "fmsb\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) ; vfnmasb, vfnmadb, wfnmasb, wfnmadb, wfnmaxb (define_insn "neg_fma4" [(set (match_operand:VF_HW 0 "register_operand" "=v") (neg:VF_HW (fma:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v") (match_operand:VF_HW 3 "register_operand" "v"))))] "TARGET_VXE" "fnmab\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) ; vfnmssb, vfnmsdb, wfnmssb, wfnmsdb, wfnmsxb (define_insn "neg_fms4" [(set (match_operand:VF_HW 0 "register_operand" "=v") (neg:VF_HW (fma:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v") (neg:VF_HW (match_operand:VF_HW 3 "register_operand" "v")))))] "TARGET_VXE" "fnmsb\t%v0,%v1,%v2,%v3" [(set_attr "op_type" "VRR")]) ; vflcsb, vflcdb, wflcsb, wflcdb, wflcxb (define_insn "neg2" [(set (match_operand:VFT 0 "register_operand" "=v") (neg:VFT (match_operand:VFT 1 "register_operand" "v")))] "TARGET_VX" "flcb\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; vflpsb, vflpdb, wflpsb, wflpdb, wflpxb (define_insn "abs2" [(set (match_operand:VFT 0 "register_operand" "=v") (abs:VFT (match_operand:VFT 1 "register_operand" "v")))] "TARGET_VX" "flpb\t%v0,%v1" [(set_attr "op_type" "VRR")]) ; vflnsb, vflndb, wflnsb, wflndb, wflnxb (define_insn "negabs2" [(set (match_operand:VFT 0 "register_operand" "=v") (neg:VFT (abs:VFT (match_operand:VFT 1 "register_operand" "v"))))] "TARGET_VX" "flnb\t%v0,%v1" [(set_attr "op_type" "VRR")]) (define_expand "smax3" [(set (match_operand:VF_HW 0 "register_operand") (smax:VF_HW (match_operand:VF_HW 1 "register_operand") (match_operand:VF_HW 2 "register_operand")))] "TARGET_VX") ; vfmaxsb, vfmaxdb, wfmaxsb, wfmaxdb, wfmaxxb (define_insn "*smax3_vxe" [(set (match_operand:VF_HW 0 "register_operand" "=v") (smax:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VXE" "fmaxb\t%v0,%v1,%v2,4" [(set_attr "op_type" "VRR")]) ; Emulate with compare + select (define_insn_and_split "*smaxv2df3_vx" [(set (match_operand:V2DF 0 "register_operand" "=v") (smax:V2DF (match_operand:V2DF 1 "register_operand" "v") (match_operand:V2DF 2 "register_operand" "v")))] "TARGET_VX && !TARGET_VXE" "#" "&& 1" [(set (match_dup 3) (not:V2DI (unge:V2DI (match_dup 2) (match_dup 1)))) (set (match_dup 0) (if_then_else:V2DF (eq (match_dup 3) (match_dup 4)) (match_dup 2) (match_dup 1)))] { operands[3] = gen_reg_rtx (V2DImode); operands[4] = CONST0_RTX (V2DImode); }) (define_expand "smin3" [(set (match_operand:VF_HW 0 "register_operand") (smin:VF_HW (match_operand:VF_HW 1 "register_operand") (match_operand:VF_HW 2 "register_operand")))] "TARGET_VX") ; vfminsb, vfmindb, wfminsb, wfmindb, wfminxb (define_insn "*smin3_vxe" [(set (match_operand:VF_HW 0 "register_operand" "=v") (smin:VF_HW (match_operand:VF_HW 1 "register_operand" "v") (match_operand:VF_HW 2 "register_operand" "v")))] "TARGET_VXE" "fminb\t%v0,%v1,%v2,4" [(set_attr "op_type" "VRR")]) ; Emulate with compare + select (define_insn_and_split "*sminv2df3_vx" [(set (match_operand:V2DF 0 "register_operand" "=v") (smin:V2DF (match_operand:V2DF 1 "register_operand" "v") (match_operand:V2DF 2 "register_operand" "v")))] "TARGET_VX && !TARGET_VXE" "#" "&& 1" [(set (match_dup 3) (not:V2DI (unge:V2DI (match_dup 2) (match_dup 1)))) (set (match_dup 0) (if_then_else:V2DF (eq (match_dup 3) (match_dup 4)) (match_dup 1) (match_dup 2)))] { operands[3] = gen_reg_rtx (V2DImode); operands[4] = CONST0_RTX (V2DImode); }) ; Vector copysign, implement using vector select (define_expand "copysign3" [(set (match_operand:VFT 0 "register_operand" "") (if_then_else:VFT (eq (match_dup 3) (match_dup 4)) (match_operand:VFT 1 "register_operand" "") (match_operand:VFT 2 "register_operand" "")))] "TARGET_VX" { int sz = GET_MODE_BITSIZE (GET_MODE_INNER (mode)); int prec = GET_MODE_PRECISION (GET_MODE_INNER (mode)); wide_int mask_val = wi::shwi (1l << (sz - 1), prec); rtx mask = gen_reg_rtx (mode); int nunits = GET_MODE_NUNITS (mode); rtvec v = rtvec_alloc (nunits); for (int i = 0; i < nunits; i++) RTVEC_ELT (v, i) = GEN_INT (mask_val.to_shwi ()); mask = gen_rtx_CONST_VECTOR (mode, v); operands[3] = force_reg (mode, mask); operands[4] = CONST0_RTX (mode); }) ;; ;; Integer compares ;; (define_insn "*vec_cmp_nocc" [(set (match_operand:VI 2 "register_operand" "=v") (VICMP_HW_OP:VI (match_operand:VI 0 "register_operand" "v") (match_operand:VI 1 "register_operand" "v")))] "TARGET_VX" "vc\t%v2,%v0,%v1" [(set_attr "op_type" "VRR")]) ;; ;; Floating point compares ;; ; vfcesb, vfcedb, wfcexb: non-signaling "==" comparison (a == b) (define_insn "*vec_cmpeq_quiet_nocc" [(set (match_operand: 0 "register_operand" "=v") (eq: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))] "TARGET_VX" "fceb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfchsb, vfchdb, wfchxb: non-signaling > comparison (!(b u>= a)) (define_insn "vec_cmpgt_quiet_nocc" [(set (match_operand: 0 "register_operand" "=v") (not: (unge: (match_operand:VFT 2 "register_operand" "v") (match_operand:VFT 1 "register_operand" "v"))))] "TARGET_VX" "fchb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) (define_expand "vec_cmplt_quiet_nocc" [(set (match_operand: 0 "register_operand" "=v") (not: (unge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v"))))] "TARGET_VX") ; vfchesb, vfchedb, wfchexb: non-signaling >= comparison (!(a u< b)) (define_insn "vec_cmpge_quiet_nocc" [(set (match_operand: 0 "register_operand" "=v") (not: (unlt: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v"))))] "TARGET_VX" "fcheb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) (define_expand "vec_cmple_quiet_nocc" [(set (match_operand: 0 "register_operand" "=v") (not: (unlt: (match_operand:VFT 2 "register_operand" "v") (match_operand:VFT 1 "register_operand" "v"))))] "TARGET_VX") ; vfkesb, vfkedb, wfkexb: signaling == comparison ((a >= b) & (b >= a)) (define_insn "*vec_cmpeq_signaling_nocc" [(set (match_operand: 0 "register_operand" "=v") (and: (ge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")) (ge: (match_dup 2) (match_dup 1))))] "TARGET_VXE" "fkeb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfkhsb, vfkhdb, wfkhxb: signaling > comparison (a > b) (define_insn "*vec_cmpgt_signaling_nocc" [(set (match_operand: 0 "register_operand" "=v") (gt: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))] "TARGET_VXE" "fkhb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) (define_insn "*vec_cmpgt_signaling_finite_nocc" [(set (match_operand: 0 "register_operand" "=v") (gt: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))] "TARGET_NONSIGNALING_VECTOR_COMPARE_OK" "fchb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; vfkhesb, vfkhedb, wfkhexb: signaling >= comparison (a >= b) (define_insn "*vec_cmpge_signaling_nocc" [(set (match_operand: 0 "register_operand" "=v") (ge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))] "TARGET_VXE" "fkheb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) (define_insn "*vec_cmpge_signaling_finite_nocc" [(set (match_operand: 0 "register_operand" "=v") (ge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))] "TARGET_NONSIGNALING_VECTOR_COMPARE_OK" "fcheb\t%v0,%v1,%v2" [(set_attr "op_type" "VRR")]) ; Expanders for not directly supported comparisons ; Signaling comparisons must be expressed via signaling rtxes only, ; and quiet comparisons must be expressed via quiet rtxes only. ; UNGT a u> b -> !!(b u< a) (define_expand "vec_cmpungt" [(set (match_operand: 0 "register_operand" "=v") (not: (unlt: (match_operand:VFT 2 "register_operand" "v") (match_operand:VFT 1 "register_operand" "v")))) (set (match_dup 0) (not: (match_dup 0)))] "TARGET_VX") ; UNGE a u>= b -> !!(a u>= b) (define_expand "vec_cmpunge" [(set (match_operand: 0 "register_operand" "=v") (not: (unge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))) (set (match_dup 0) (not: (match_dup 0)))] "TARGET_VX") ; UNEQ a u== b -> !(!(a u>= b) | !(b u>= a)) (define_expand "vec_cmpuneq" [(set (match_operand: 0 "register_operand" "=v") (not: (unge: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))) (set (match_dup 3) (not: (unge: (match_dup 2) (match_dup 1)))) (set (match_dup 0) (ior: (match_dup 0) (match_dup 3))) (set (match_dup 0) (not: (match_dup 0)))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); }) ; LTGT a <> b -> a > b | b > a (define_expand "vec_cmpltgt" [(set (match_operand: 0 "register_operand" "=v") (gt: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v"))) (set (match_dup 3) (gt: (match_dup 2) (match_dup 1))) (set (match_dup 0) (ior: (match_dup 0) (match_dup 3)))] "TARGET_VXE" { operands[3] = gen_reg_rtx (mode); }) ; ORDERED (a, b): !(a u< b) | !(a u>= b) (define_expand "vec_cmpordered" [(set (match_operand: 0 "register_operand" "=v") (not: (unlt: (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")))) (set (match_dup 3) (not: (unge: (match_dup 1) (match_dup 2)))) (set (match_dup 0) (ior: (match_dup 0) (match_dup 3)))] "TARGET_VX" { operands[3] = gen_reg_rtx (mode); }) ; UNORDERED (a, b): !ORDERED (a, b) (define_expand "vec_cmpunordered" [(match_operand: 0 "register_operand" "=v") (match_operand:VFT 1 "register_operand" "v") (match_operand:VFT 2 "register_operand" "v")] "TARGET_VX" { emit_insn (gen_vec_cmpordered (operands[0], operands[1], operands[2])); emit_insn (gen_rtx_SET (operands[0], gen_rtx_NOT (mode, operands[0]))); DONE; }) (define_code_iterator VEC_CMP_EXPAND [ungt unge uneq ltgt ordered unordered]) (define_expand "vec_cmp" [(match_operand 0 "register_operand" "") (VEC_CMP_EXPAND (match_operand 1 "register_operand" "") (match_operand 2 "register_operand" ""))] "TARGET_VX" { if (GET_MODE (operands[1]) == V4SFmode) emit_insn (gen_vec_cmpv4sf (operands[0], operands[1], operands[2])); else if (GET_MODE (operands[1]) == V2DFmode) emit_insn (gen_vec_cmpv2df (operands[0], operands[1], operands[2])); else gcc_unreachable (); DONE; }) (define_insn "*vec_load_pair" [(set (match_operand:V_HW_64 0 "register_operand" "=v,v") (vec_concat:V_HW_64 (match_operand: 1 "register_operand" "d,v") (match_operand: 2 "register_operand" "d,v")))] "TARGET_VX" "@ vlvgp\t%v0,%1,%2 vmrhg\t%v0,%v1,%v2" [(set_attr "op_type" "VRR,VRR")]) (define_insn "vllv16qi" [(set (match_operand:V16QI 0 "register_operand" "=v") (unspec:V16QI [(match_operand:SI 1 "register_operand" "d") (match_operand:BLK 2 "memory_operand" "Q")] UNSPEC_VEC_LOAD_LEN))] "TARGET_VX" "vll\t%v0,%1,%2" [(set_attr "op_type" "VRS")]) ; vfenebs, vfenehs, vfenefs ; vfenezbs, vfenezhs, vfenezfs (define_insn "vec_vfenes" [(set (match_operand:VI_HW_QHS 0 "register_operand" "=v") (unspec:VI_HW_QHS [(match_operand:VI_HW_QHS 1 "register_operand" "v") (match_operand:VI_HW_QHS 2 "register_operand" "v") (match_operand:QI 3 "const_mask_operand" "C")] UNSPEC_VEC_VFENE)) (set (reg:CCRAW CC_REGNUM) (unspec:CCRAW [(match_dup 1) (match_dup 2) (match_dup 3)] UNSPEC_VEC_VFENECC))] "TARGET_VX" { unsigned HOST_WIDE_INT flags = UINTVAL (operands[3]); gcc_assert (!(flags & ~(VSTRING_FLAG_ZS | VSTRING_FLAG_CS))); flags &= ~VSTRING_FLAG_CS; if (flags == VSTRING_FLAG_ZS) return "vfenezs\t%v0,%v1,%v2"; return "vfenes\t%v0,%v1,%v2"; } [(set_attr "op_type" "VRR")]) ; Vector select ; The following splitters simplify vec_sel for constant 0 or -1 ; selection sources. This is required to generate efficient code for ; vcond. ; a = b == c; (define_split [(set (match_operand:V 0 "register_operand" "") (if_then_else:V (eq (match_operand: 3 "register_operand" "") (match_operand:V 4 "const0_operand" "")) (match_operand:V 1 "const0_operand" "") (match_operand:V 2 "all_ones_operand" "")))] "TARGET_VX" [(set (match_dup 0) (match_dup 3))] { PUT_MODE (operands[3], mode); }) ; a = ~(b == c) (define_split [(set (match_operand:V 0 "register_operand" "") (if_then_else:V (eq (match_operand: 3 "register_operand" "") (match_operand:V 4 "const0_operand" "")) (match_operand:V 1 "all_ones_operand" "") (match_operand:V 2 "const0_operand" "")))] "TARGET_VX" [(set (match_dup 0) (not:V (match_dup 3)))] { PUT_MODE (operands[3], mode); }) ; a = b != c (define_split [(set (match_operand:V 0 "register_operand" "") (if_then_else:V (ne (match_operand: 3 "register_operand" "") (match_operand:V 4 "const0_operand" "")) (match_operand:V 1 "all_ones_operand" "") (match_operand:V 2 "const0_operand" "")))] "TARGET_VX" [(set (match_dup 0) (match_dup 3))] { PUT_MODE (operands[3], mode); }) ; a = ~(b != c) (define_split [(set (match_operand:V 0 "register_operand" "") (if_then_else:V (ne (match_operand: 3 "register_operand" "") (match_operand:V 4 "const0_operand" "")) (match_operand:V 1 "const0_operand" "") (match_operand:V 2 "all_ones_operand" "")))] "TARGET_VX" [(set (match_dup 0) (not:V (match_dup 3)))] { PUT_MODE (operands[3], mode); }) ; op0 = op3 == 0 ? op1 : op2 (define_insn "*vec_sel0" [(set (match_operand:V 0 "register_operand" "=v") (if_then_else:V (eq (match_operand: 3 "register_operand" "v") (match_operand: 4 "const0_operand" "")) (match_operand:V 1 "register_operand" "v") (match_operand:V 2 "register_operand" "v")))] "TARGET_VX" "vsel\t%v0,%2,%1,%3" [(set_attr "op_type" "VRR")]) ; op0 = !op3 == 0 ? op1 : op2 (define_insn "*vec_sel0" [(set (match_operand:V 0 "register_operand" "=v") (if_then_else:V (eq (not: (match_operand: 3 "register_operand" "v")) (match_operand: 4 "const0_operand" "")) (match_operand:V 1 "register_operand" "v") (match_operand:V 2 "register_operand" "v")))] "TARGET_VX" "vsel\t%v0,%1,%2,%3" [(set_attr "op_type" "VRR")]) ; op0 = op3 == -1 ? op1 : op2 (define_insn "*vec_sel1" [(set (match_operand:V 0 "register_operand" "=v") (if_then_else:V (eq (match_operand: 3 "register_operand" "v") (match_operand: 4 "all_ones_operand" "")) (match_operand:V 1 "register_operand" "v") (match_operand:V 2 "register_operand" "v")))] "TARGET_VX" "vsel\t%v0,%1,%2,%3" [(set_attr "op_type" "VRR")]) ; op0 = !op3 == -1 ? op1 : op2 (define_insn "*vec_sel1" [(set (match_operand:V 0 "register_operand" "=v") (if_then_else:V (eq (not: (match_operand: 3 "register_operand" "v")) (match_operand: 4 "all_ones_operand" "")) (match_operand:V 1 "register_operand" "v") (match_operand:V 2 "register_operand" "v")))] "TARGET_VX" "vsel\t%v0,%2,%1,%3" [(set_attr "op_type" "VRR")]) ; vec_pack_trunc ; vpkh, vpkf, vpkg (define_insn "vec_pack_trunc_" [(set (match_operand: 0 "register_operand" "=v") (vec_concat: (truncate: (match_operand:VI_HW_HSD 1 "register_operand" "v")) (truncate: (match_operand:VI_HW_HSD 2 "register_operand" "v"))))] "TARGET_VX" "vpk\t%0,%1,%2" [(set_attr "op_type" "VRR")]) ; vpksh, vpksf, vpksg (define_insn "vec_pack_ssat_" [(set (match_operand: 0 "register_operand" "=v") (vec_concat: (ss_truncate: (match_operand:VI_HW_HSD 1 "register_operand" "v")) (ss_truncate: (match_operand:VI_HW_HSD 2 "register_operand" "v"))))] "TARGET_VX" "vpks\t%0,%1,%2" [(set_attr "op_type" "VRR")]) ; vpklsh, vpklsf, vpklsg (define_insn "vec_pack_usat_" [(set (match_operand: 0 "register_operand" "=v") (vec_concat: (us_truncate: (match_operand:VI_HW_HSD 1 "register_operand" "v")) (us_truncate: (match_operand:VI_HW_HSD 2 "register_operand" "v"))))] "TARGET_VX" "vpkls\t%0,%1,%2" [(set_attr "op_type" "VRR")]) ;; vector unpack v16qi ; signed (define_insn "vec_unpacks_hi_v16qi" [(set (match_operand:V8HI 0 "register_operand" "=v") (sign_extend:V8HI (vec_select:V8QI (match_operand:V16QI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)(const_int 2)(const_int 3) (const_int 4)(const_int 5)(const_int 6)(const_int 7)]))))] "TARGET_VX" "vuphb\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacks_lo_v16qi" [(set (match_operand:V8HI 0 "register_operand" "=v") (sign_extend:V8HI (vec_select:V8QI (match_operand:V16QI 1 "register_operand" "v") (parallel [(const_int 8) (const_int 9) (const_int 10)(const_int 11) (const_int 12)(const_int 13)(const_int 14)(const_int 15)]))))] "TARGET_VX" "vuplb\t%0,%1" [(set_attr "op_type" "VRR")]) ; unsigned (define_insn "vec_unpacku_hi_v16qi" [(set (match_operand:V8HI 0 "register_operand" "=v") (zero_extend:V8HI (vec_select:V8QI (match_operand:V16QI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)(const_int 2)(const_int 3) (const_int 4)(const_int 5)(const_int 6)(const_int 7)]))))] "TARGET_VX" "vuplhb\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacku_lo_v16qi" [(set (match_operand:V8HI 0 "register_operand" "=v") (zero_extend:V8HI (vec_select:V8QI (match_operand:V16QI 1 "register_operand" "v") (parallel [(const_int 8) (const_int 9) (const_int 10)(const_int 11) (const_int 12)(const_int 13)(const_int 14)(const_int 15)]))))] "TARGET_VX" "vupllb\t%0,%1" [(set_attr "op_type" "VRR")]) ;; vector unpack v8hi ; signed (define_insn "vec_unpacks_hi_v8hi" [(set (match_operand:V4SI 0 "register_operand" "=v") (sign_extend:V4SI (vec_select:V4HI (match_operand:V8HI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)(const_int 2)(const_int 3)]))))] "TARGET_VX" "vuphh\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacks_lo_v8hi" [(set (match_operand:V4SI 0 "register_operand" "=v") (sign_extend:V4SI (vec_select:V4HI (match_operand:V8HI 1 "register_operand" "v") (parallel [(const_int 4)(const_int 5)(const_int 6)(const_int 7)]))))] "TARGET_VX" "vuplhw\t%0,%1" [(set_attr "op_type" "VRR")]) ; unsigned (define_insn "vec_unpacku_hi_v8hi" [(set (match_operand:V4SI 0 "register_operand" "=v") (zero_extend:V4SI (vec_select:V4HI (match_operand:V8HI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)(const_int 2)(const_int 3)]))))] "TARGET_VX" "vuplhh\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacku_lo_v8hi" [(set (match_operand:V4SI 0 "register_operand" "=v") (zero_extend:V4SI (vec_select:V4HI (match_operand:V8HI 1 "register_operand" "v") (parallel [(const_int 4)(const_int 5)(const_int 6)(const_int 7)]))))] "TARGET_VX" "vupllh\t%0,%1" [(set_attr "op_type" "VRR")]) ;; vector unpack v4si ; signed (define_insn "vec_unpacks_hi_v4si" [(set (match_operand:V2DI 0 "register_operand" "=v") (sign_extend:V2DI (vec_select:V2SI (match_operand:V4SI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)]))))] "TARGET_VX" "vuphf\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacks_lo_v4si" [(set (match_operand:V2DI 0 "register_operand" "=v") (sign_extend:V2DI (vec_select:V2SI (match_operand:V4SI 1 "register_operand" "v") (parallel [(const_int 2)(const_int 3)]))))] "TARGET_VX" "vuplf\t%0,%1" [(set_attr "op_type" "VRR")]) ; unsigned (define_insn "vec_unpacku_hi_v4si" [(set (match_operand:V2DI 0 "register_operand" "=v") (zero_extend:V2DI (vec_select:V2SI (match_operand:V4SI 1 "register_operand" "v") (parallel [(const_int 0)(const_int 1)]))))] "TARGET_VX" "vuplhf\t%0,%1" [(set_attr "op_type" "VRR")]) (define_insn "vec_unpacku_lo_v4si" [(set (match_operand:V2DI 0 "register_operand" "=v") (zero_extend:V2DI (vec_select:V2SI (match_operand:V4SI 1 "register_operand" "v") (parallel [(const_int 2)(const_int 3)]))))] "TARGET_VX" "vupllf\t%0,%1" [(set_attr "op_type" "VRR")]) ;; vector load lengthened ; vflls float -> double (define_insn "*vec_extendv4sf" [(set (match_operand:V2DF 0 "register_operand" "=v") (float_extend:V2DF (vec_select:V2SF (match_operand:V4SF 1 "register_operand" "v") (parallel [(const_int 0) (const_int 2)]))))] "TARGET_VX" "vldeb\t%v0,%v1" [(set_attr "op_type" "VRR")]) (define_expand "vec_unpacks_lo_v4sf" [(set (match_dup 2) (unspec:V4SF [(match_operand:V4SF 1 "register_operand" "v") (match_dup 1)] UNSPEC_VEC_MERGEL)) (set (match_operand:V2DF 0 "register_operand" "=v") (float_extend:V2DF (vec_select:V2SF (match_dup 2) (parallel [(const_int 0) (const_int 2)]))))] "TARGET_VX" { operands[2] = gen_reg_rtx(V4SFmode); }) (define_expand "vec_unpacks_hi_v4sf" [(set (match_dup 2) (unspec:V4SF [(match_operand:V4SF 1 "register_operand" "v") (match_dup 1)] UNSPEC_VEC_MERGEH)) (set (match_operand:V2DF 0 "register_operand" "=v") (float_extend:V2DF (vec_select:V2SF (match_dup 2) (parallel [(const_int 0) (const_int 2)]))))] "TARGET_VX" { operands[2] = gen_reg_rtx(V4SFmode); }) ; double -> long double (define_insn "*vec_extendv2df" [(set (match_operand:V1TF 0 "register_operand" "=v") (float_extend:V1TF (vec_select:V1DF (match_operand:V2DF 1 "register_operand" "v") (parallel [(const_int 0)]))))] "TARGET_VXE" "wflld\t%v0,%v1" [(set_attr "op_type" "VRR")]) (define_expand "vec_unpacks_lo_v2df" [(set (match_dup 2) (unspec:V2DF [(match_operand:V2DF 1 "register_operand" "v") (match_dup 1)] UNSPEC_VEC_MERGEL)) (set (match_operand:V1TF 0 "register_operand" "=v") (float_extend:V1TF (vec_select:V1DF (match_dup 2) (parallel [(const_int 0)]))))] "TARGET_VXE" { operands[2] = gen_reg_rtx (V2DFmode); }) (define_expand "vec_unpacks_hi_v2df" [(set (match_dup 2) (unspec:V2DF [(match_operand:V2DF 1 "register_operand" "v") (match_dup 1)] UNSPEC_VEC_MERGEH)) (set (match_operand:V1TF 0 "register_operand" "=v") (float_extend:V1TF (vec_select:V1DF (match_dup 2) (parallel [(const_int 0)]))))] "TARGET_VXE" { operands[2] = gen_reg_rtx (V2DFmode); }) ; 2 x v2df -> 1 x v4sf (define_expand "vec_pack_trunc_v2df" [(set (match_dup 3) (unspec:V4SF [(match_operand:V2DF 1 "register_operand" "") (const_int VEC_INEXACT) (const_int VEC_RND_CURRENT)] UNSPEC_VEC_VFLR)) (set (match_dup 4) (unspec:V4SF [(match_operand:V2DF 2 "register_operand" "") (const_int VEC_INEXACT) (const_int VEC_RND_CURRENT)] UNSPEC_VEC_VFLR)) (set (match_dup 6) (unspec:V16QI [(subreg:V16QI (match_dup 3) 0) (subreg:V16QI (match_dup 4) 0) (match_dup 5)] UNSPEC_VEC_PERM)) (set (match_operand:V4SF 0 "register_operand" "") (subreg:V4SF (match_dup 6) 0))] "TARGET_VX" { rtx constv, perm[16]; int i; for (i = 0; i < 4; ++i) { perm[i] = GEN_INT (i); perm[i + 4] = GEN_INT (i + 8); perm[i + 8] = GEN_INT (i + 16); perm[i + 12] = GEN_INT (i + 24); } constv = gen_rtx_CONST_VECTOR (V16QImode, gen_rtvec_v (16, perm)); operands[3] = gen_reg_rtx (V4SFmode); operands[4] = gen_reg_rtx (V4SFmode); operands[5] = force_reg (V16QImode, constv); operands[6] = gen_reg_rtx (V16QImode); }) ; ; BFP <-> integer conversions ; ; signed integer to floating point ; op2: inexact exception not suppressed (IEEE 754 2008) ; op3: according to current rounding mode ; vcdgb, vcefb (define_insn "float2" [(set (match_operand:VX_VEC_CONV_BFP 0 "register_operand" "=v") (float:VX_VEC_CONV_BFP (match_operand:VX_VEC_CONV_INT 1 "register_operand" "v")))] "TARGET_VX && GET_MODE_UNIT_SIZE (mode) == GET_MODE_UNIT_SIZE (mode)" "vcb\t%v0,%v1,0,0" [(set_attr "op_type" "VRR")]) ; unsigned integer to floating point ; op2: inexact exception not suppressed (IEEE 754 2008) ; op3: according to current rounding mode ; vcdlgb, vcelfb (define_insn "floatuns2" [(set (match_operand:VX_VEC_CONV_BFP 0 "register_operand" "=v") (unsigned_float:VX_VEC_CONV_BFP (match_operand:VX_VEC_CONV_INT 1 "register_operand" "v")))] "TARGET_VX && GET_MODE_UNIT_SIZE (mode) == GET_MODE_UNIT_SIZE (mode)" "vclb\t%v0,%v1,0,0" [(set_attr "op_type" "VRR")]) ; floating point to signed integer ; op2: inexact exception not suppressed (IEEE 754 2008) ; op3: rounding mode 5 (round towards 0 C11 6.3.1.4) ; vcgdb, vcfeb (define_insn "fix_trunc2" [(set (match_operand:VX_VEC_CONV_INT 0 "register_operand" "=v") (fix:VX_VEC_CONV_INT (match_operand:VX_VEC_CONV_BFP 1 "register_operand" "v")))] "TARGET_VX && GET_MODE_UNIT_SIZE (mode) == GET_MODE_UNIT_SIZE (mode)" "vcb\t%v0,%v1,0,5" [(set_attr "op_type" "VRR")]) ; floating point to unsigned integer ; op2: inexact exception not suppressed (IEEE 754 2008) ; op3: rounding mode 5 (round towards 0 C11 6.3.1.4) ; vclgdb, vclfeb (define_insn "fixuns_trunc2" [(set (match_operand:VX_VEC_CONV_INT 0 "register_operand" "=v") (unsigned_fix:VX_VEC_CONV_INT (match_operand:VX_VEC_CONV_BFP 1 "register_operand" "v")))] "TARGET_VX && GET_MODE_UNIT_SIZE (mode) == GET_MODE_UNIT_SIZE (mode)" "vclb\t%v0,%v1,0,5" [(set_attr "op_type" "VRR")]) ; ; Vector byte swap patterns ; ; FIXME: The bswap rtl standard name currently does not appear to be ; used for vector modes. (define_expand "bswap" [(parallel [(set (match_operand:VT_HW_HSDT 0 "nonimmediate_operand" "") (bswap:VT_HW_HSDT (match_operand:VT_HW_HSDT 1 "nonimmediate_operand" ""))) (use (match_dup 2))])] "TARGET_VX" { static char p[4][16] = { { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 }, /* H */ { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 }, /* S */ { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 }, /* D */ { 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0 } }; /* T */ char *perm; rtx perm_rtx[16]; switch (GET_MODE_SIZE (GET_MODE_INNER (mode))) { case 2: perm = p[0]; break; case 4: perm = p[1]; break; case 8: perm = p[2]; break; case 16: perm = p[3]; break; default: gcc_unreachable (); } for (int i = 0; i < 16; i++) perm_rtx[i] = GEN_INT (perm[i]); operands[2] = gen_rtx_CONST_VECTOR (V16QImode, gen_rtvec_v (16, perm_rtx)); /* Without vxe2 we do not have byte swap instructions dealing directly with memory operands. So instead of waiting until reload to fix that up switch over to vector permute right now. */ if (!TARGET_VXE2) { rtx in = force_reg (V16QImode, simplify_gen_subreg (V16QImode, operands[1], mode, 0)); rtx permute = force_reg (V16QImode, force_const_mem (V16QImode, operands[2])); rtx out = gen_reg_rtx (V16QImode); emit_insn (gen_vec_permv16qi (out, in, in, permute)); emit_move_insn (operands[0], simplify_gen_subreg (mode, out, V16QImode, 0)); DONE; } }) ; Switching late to the reg-reg variant requires the vector permute ; pattern to be pushed into literal pool and allocating a vector ; register to load it into. We rely on both being provided by LRA ; when fixing up the v constraint for operand 2. ; permute_pattern_operand: general_operand would reject the permute ; pattern constants since these are not accepted by ; s390_legimitate_constant_p ; ^R: Prevent these alternatives from being chosen if it would require ; pushing the operand into memory first ; vlbrh, vlbrf, vlbrg, vlbrq, vstbrh, vstbrf, vstbrg, vstbrq (define_insn_and_split "*bswap" [(set (match_operand:VT_HW_HSDT 0 "nonimmediate_operand" "=v, v,^R") (bswap:VT_HW_HSDT (match_operand:VT_HW_HSDT 1 "nonimmediate_operand" "v,^R, v"))) (use (match_operand:V16QI 2 "permute_pattern_operand" "v, X, X"))] "TARGET_VXE2" "@ # vlbr\t%v0,%v1 vstbr\t%v1,%v0" "&& reload_completed && !memory_operand (operands[0], mode) && !memory_operand (operands[1], mode)" [(set (match_dup 0) (subreg:VT_HW_HSDT (unspec:V16QI [(subreg:V16QI (match_dup 1) 0) (subreg:V16QI (match_dup 1) 0) (match_dup 2)] UNSPEC_VEC_PERM) 0))] "" [(set_attr "op_type" "*,VRX,VRX")]) ; reduc_smin ; reduc_smax ; reduc_umin ; reduc_umax ; vec_pack_sfix_trunc: convert + pack ? ; vec_pack_ufix_trunc ; vec_unpacks_float_hi ; vec_unpacks_float_lo ; vec_unpacku_float_hi ; vec_unpacku_float_lo