//===- SystemZInstrInfo.td - SystemZ Instruction defs ---------*- tblgen-*-===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// // // This file describes the SystemZ instructions in TableGen format. // //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // SystemZ Instruction Predicate Definitions. def IsZ10 : Predicate<"Subtarget.isZ10()">; include "SystemZInstrFormats.td" //===----------------------------------------------------------------------===// // Type Constraints. //===----------------------------------------------------------------------===// class SDTCisI8 : SDTCisVT; class SDTCisI16 : SDTCisVT; class SDTCisI32 : SDTCisVT; class SDTCisI64 : SDTCisVT; //===----------------------------------------------------------------------===// // Type Profiles. //===----------------------------------------------------------------------===// def SDT_SystemZCall : SDTypeProfile<0, -1, [SDTCisPtrTy<0>]>; def SDT_SystemZCallSeqStart : SDCallSeqStart<[SDTCisI64<0>]>; def SDT_SystemZCallSeqEnd : SDCallSeqEnd<[SDTCisI64<0>, SDTCisI64<1>]>; def SDT_CmpTest : SDTypeProfile<0, 2, [SDTCisSameAs<0, 1>]>; def SDT_BrCond : SDTypeProfile<0, 2, [SDTCisVT<0, OtherVT>, SDTCisI8<1>]>; def SDT_SelectCC : SDTypeProfile<1, 3, [SDTCisSameAs<0, 1>, SDTCisSameAs<1, 2>, SDTCisI8<3>]>; def SDT_Address : SDTypeProfile<1, 1, [SDTCisSameAs<0, 1>, SDTCisPtrTy<0>]>; //===----------------------------------------------------------------------===// // SystemZ Specific Node Definitions. //===----------------------------------------------------------------------===// def SystemZretflag : SDNode<"SystemZISD::RET_FLAG", SDTNone, [SDNPHasChain, SDNPOptInFlag]>; def SystemZcall : SDNode<"SystemZISD::CALL", SDT_SystemZCall, [SDNPHasChain, SDNPOutFlag, SDNPOptInFlag]>; def SystemZcallseq_start : SDNode<"ISD::CALLSEQ_START", SDT_SystemZCallSeqStart, [SDNPHasChain, SDNPOutFlag]>; def SystemZcallseq_end : SDNode<"ISD::CALLSEQ_END", SDT_SystemZCallSeqEnd, [SDNPHasChain, SDNPOptInFlag, SDNPOutFlag]>; def SystemZcmp : SDNode<"SystemZISD::CMP", SDT_CmpTest, [SDNPOutFlag]>; def SystemZucmp : SDNode<"SystemZISD::UCMP", SDT_CmpTest, [SDNPOutFlag]>; def SystemZbrcond : SDNode<"SystemZISD::BRCOND", SDT_BrCond, [SDNPHasChain, SDNPInFlag]>; def SystemZselect : SDNode<"SystemZISD::SELECT", SDT_SelectCC, [SDNPInFlag]>; def SystemZpcrelwrapper : SDNode<"SystemZISD::PCRelativeWrapper", SDT_Address, []>; include "SystemZOperands.td" //===----------------------------------------------------------------------===// // Instruction list.. def ADJCALLSTACKDOWN : Pseudo<(outs), (ins i64imm:$amt), "#ADJCALLSTACKDOWN", [(SystemZcallseq_start timm:$amt)]>; def ADJCALLSTACKUP : Pseudo<(outs), (ins i64imm:$amt1, i64imm:$amt2), "#ADJCALLSTACKUP", [(SystemZcallseq_end timm:$amt1, timm:$amt2)]>; let usesCustomDAGSchedInserter = 1 in { def Select32 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2, i8imm:$cc), "# Select32 PSEUDO", [(set GR32:$dst, (SystemZselect GR32:$src1, GR32:$src2, imm:$cc))]>; def Select64 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2, i8imm:$cc), "# Select64 PSEUDO", [(set GR64:$dst, (SystemZselect GR64:$src1, GR64:$src2, imm:$cc))]>; } //===----------------------------------------------------------------------===// // Control Flow Instructions... // // FIXME: Provide proper encoding! let isReturn = 1, isTerminator = 1, isBarrier = 1, hasCtrlDep = 1 in { def RET : Pseudo<(outs), (ins), "br\t%r14", [(SystemZretflag)]>; } let isBranch = 1, isTerminator = 1 in { let isBarrier = 1 in { def JMP : Pseudo<(outs), (ins brtarget:$dst), "j\t{$dst}", [(br bb:$dst)]>; let isIndirectBranch = 1 in def JMPr : Pseudo<(outs), (ins GR64:$dst), "br\t{$dst}", [(brind GR64:$dst)]>; } let Uses = [PSW] in { def JO : Pseudo<(outs), (ins brtarget:$dst), "jo\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_O)]>; def JH : Pseudo<(outs), (ins brtarget:$dst), "jh\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_H)]>; def JNLE: Pseudo<(outs), (ins brtarget:$dst), "jnle\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NLE)]>; def JL : Pseudo<(outs), (ins brtarget:$dst), "jl\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_L)]>; def JNHE: Pseudo<(outs), (ins brtarget:$dst), "jnhe\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NHE)]>; def JLH : Pseudo<(outs), (ins brtarget:$dst), "jlh\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_LH)]>; def JNE : Pseudo<(outs), (ins brtarget:$dst), "jne\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NE)]>; def JE : Pseudo<(outs), (ins brtarget:$dst), "je\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_E)]>; def JNLH: Pseudo<(outs), (ins brtarget:$dst), "jnlh\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NLH)]>; def JHE : Pseudo<(outs), (ins brtarget:$dst), "jhe\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_HE)]>; def JNL : Pseudo<(outs), (ins brtarget:$dst), "jnl\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NL)]>; def JLE : Pseudo<(outs), (ins brtarget:$dst), "jle\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_LE)]>; def JNH : Pseudo<(outs), (ins brtarget:$dst), "jnh\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NH)]>; def JNO : Pseudo<(outs), (ins brtarget:$dst), "jno\t$dst", [(SystemZbrcond bb:$dst, SYSTEMZ_COND_NO)]>; } // Uses = [PSW] } // isBranch = 1 //===----------------------------------------------------------------------===// // Call Instructions... // let isCall = 1 in // All calls clobber the non-callee saved registers. Uses for argument // registers are added manually. let Defs = [R0D, R1D, R2D, R3D, R4D, R5D, R14D, F0L, F1L, F2L, F3L, F4L, F5L, F6L, F7L] in { def CALLi : Pseudo<(outs), (ins imm_pcrel:$dst, variable_ops), "brasl\t%r14, $dst", [(SystemZcall imm:$dst)]>; def CALLr : Pseudo<(outs), (ins ADDR64:$dst, variable_ops), "basr\t%r14, $dst", [(SystemZcall ADDR64:$dst)]>; } //===----------------------------------------------------------------------===// // Miscellaneous Instructions. // let isReMaterializable = 1 in // FIXME: Provide imm12 variant // FIXME: Address should be halfword aligned... def LA64r : Pseudo<(outs GR64:$dst), (ins laaddr:$src), "lay\t{$dst, $src}", [(set GR64:$dst, laaddr:$src)]>; def LA64rm : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "larl\t{$dst, $src}", [(set GR64:$dst, (SystemZpcrelwrapper tglobaladdr:$src))]>; let neverHasSideEffects = 1 in def NOP : Pseudo<(outs), (ins), "# no-op", []>; //===----------------------------------------------------------------------===// // Move Instructions // FIXME: Provide proper encoding! let neverHasSideEffects = 1 in { def MOV32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src), "lr\t{$dst, $src}", []>; def MOV64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src), "lgr\t{$dst, $src}", []>; def MOV128rr : Pseudo<(outs GR128:$dst), (ins GR128:$src), "# MOV128 PSEUDO!\n" "\tlgr\t${dst:subreg_odd}, ${src:subreg_odd}\n" "\tlgr\t${dst:subreg_even}, ${src:subreg_even}", []>; def MOV64rrP : Pseudo<(outs GR64P:$dst), (ins GR64P:$src), "# MOV64P PSEUDO!\n" "\tlr\t${dst:subreg_odd}, ${src:subreg_odd}\n" "\tlr\t${dst:subreg_even}, ${src:subreg_even}", []>; } def MOVSX64rr32 : Pseudo<(outs GR64:$dst), (ins GR32:$src), "lgfr\t{$dst, $src}", [(set GR64:$dst, (sext GR32:$src))]>; def MOVZX64rr32 : Pseudo<(outs GR64:$dst), (ins GR32:$src), "llgfr\t{$dst, $src}", [(set GR64:$dst, (zext GR32:$src))]>; // FIXME: Provide proper encoding! let isReMaterializable = 1, isAsCheapAsAMove = 1 in { def MOV32ri16 : Pseudo<(outs GR32:$dst), (ins s16imm:$src), "lhi\t{$dst, $src}", [(set GR32:$dst, immSExt16:$src)]>; def MOV64ri16 : Pseudo<(outs GR64:$dst), (ins s16imm64:$src), "lghi\t{$dst, $src}", [(set GR64:$dst, immSExt16:$src)]>; def MOV64rill16 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llill\t{$dst, $src}", [(set GR64:$dst, i64ll16:$src)]>; def MOV64rilh16 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llilh\t{$dst, $src}", [(set GR64:$dst, i64lh16:$src)]>; def MOV64rihl16 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llihl\t{$dst, $src}", [(set GR64:$dst, i64hl16:$src)]>; def MOV64rihh16 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llihh\t{$dst, $src}", [(set GR64:$dst, i64hh16:$src)]>; def MOV64ri32 : Pseudo<(outs GR64:$dst), (ins s32imm64:$src), "lgfi\t{$dst, $src}", [(set GR64:$dst, immSExt32:$src)]>; def MOV64rilo32 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llilf\t{$dst, $src}", [(set GR64:$dst, i64lo32:$src)]>; def MOV64rihi32 : Pseudo<(outs GR64:$dst), (ins i64imm:$src), "llihf\t{$dst, $src}", [(set GR64:$dst, i64hi32:$src)]>; } let canFoldAsLoad = 1, isReMaterializable = 1, mayHaveSideEffects = 1 in { def MOV32rm : Pseudo<(outs GR32:$dst), (ins rriaddr12:$src), "l\t{$dst, $src}", [(set GR32:$dst, (load rriaddr12:$src))]>; def MOV32rmy : Pseudo<(outs GR32:$dst), (ins rriaddr:$src), "ly\t{$dst, $src}", [(set GR32:$dst, (load rriaddr:$src))]>; def MOV64rm : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "lg\t{$dst, $src}", [(set GR64:$dst, (load rriaddr:$src))]>; } def MOV32mr : Pseudo<(outs), (ins rriaddr12:$dst, GR32:$src), "st\t{$src, $dst}", [(store GR32:$src, rriaddr12:$dst)]>; def MOV32mry : Pseudo<(outs), (ins rriaddr:$dst, GR32:$src), "sty\t{$src, $dst}", [(store GR32:$src, rriaddr:$dst)]>; def MOV64mr : Pseudo<(outs), (ins rriaddr:$dst, GR64:$src), "stg\t{$src, $dst}", [(store GR64:$src, rriaddr:$dst)]>; def MOV8mi : Pseudo<(outs), (ins riaddr12:$dst, i32i8imm:$src), "mvi\t{$dst, $src}", [(truncstorei8 (i32 i32immSExt8:$src), riaddr12:$dst)]>; def MOV8miy : Pseudo<(outs), (ins riaddr:$dst, i32i8imm:$src), "mviy\t{$dst, $src}", [(truncstorei8 (i32 i32immSExt8:$src), riaddr:$dst)]>; def MOV16mi : Pseudo<(outs), (ins riaddr12:$dst, s16imm:$src), "mvhhi\t{$dst, $src}", [(truncstorei16 (i32 i32immSExt16:$src), riaddr12:$dst)]>, Requires<[IsZ10]>; def MOV32mi16 : Pseudo<(outs), (ins riaddr12:$dst, s32imm:$src), "mvhi\t{$dst, $src}", [(store (i32 immSExt16:$src), riaddr12:$dst)]>, Requires<[IsZ10]>; def MOV64mi16 : Pseudo<(outs), (ins riaddr12:$dst, s32imm64:$src), "mvghi\t{$dst, $src}", [(store (i64 immSExt16:$src), riaddr12:$dst)]>, Requires<[IsZ10]>; // sexts def MOVSX32rr8 : Pseudo<(outs GR32:$dst), (ins GR32:$src), "lbr\t{$dst, $src}", [(set GR32:$dst, (sext_inreg GR32:$src, i8))]>; def MOVSX64rr8 : Pseudo<(outs GR64:$dst), (ins GR64:$src), "lgbr\t{$dst, $src}", [(set GR64:$dst, (sext_inreg GR64:$src, i8))]>; def MOVSX32rr16 : Pseudo<(outs GR32:$dst), (ins GR32:$src), "lhr\t{$dst, $src}", [(set GR32:$dst, (sext_inreg GR32:$src, i16))]>; def MOVSX64rr16 : Pseudo<(outs GR64:$dst), (ins GR64:$src), "lghr\t{$dst, $src}", [(set GR64:$dst, (sext_inreg GR64:$src, i16))]>; // extloads def MOVSX32rm8 : Pseudo<(outs GR32:$dst), (ins rriaddr:$src), "lb\t{$dst, $src}", [(set GR32:$dst, (sextloadi32i8 rriaddr:$src))]>; def MOVSX32rm16 : Pseudo<(outs GR32:$dst), (ins rriaddr12:$src), "lh\t{$dst, $src}", [(set GR32:$dst, (sextloadi32i16 rriaddr12:$src))]>; def MOVSX32rm16y : Pseudo<(outs GR32:$dst), (ins rriaddr:$src), "lhy\t{$dst, $src}", [(set GR32:$dst, (sextloadi32i16 rriaddr:$src))]>; def MOVSX64rm8 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "lgb\t{$dst, $src}", [(set GR64:$dst, (sextloadi64i8 rriaddr:$src))]>; def MOVSX64rm16 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "lgh\t{$dst, $src}", [(set GR64:$dst, (sextloadi64i16 rriaddr:$src))]>; def MOVSX64rm32 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "lgf\t{$dst, $src}", [(set GR64:$dst, (sextloadi64i32 rriaddr:$src))]>; def MOVZX32rm8 : Pseudo<(outs GR32:$dst), (ins rriaddr:$src), "llc\t{$dst, $src}", [(set GR32:$dst, (zextloadi32i8 rriaddr:$src))]>; def MOVZX32rm16 : Pseudo<(outs GR32:$dst), (ins rriaddr:$src), "llh\t{$dst, $src}", [(set GR32:$dst, (zextloadi32i16 rriaddr:$src))]>; def MOVZX64rm8 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "llgc\t{$dst, $src}", [(set GR64:$dst, (zextloadi64i8 rriaddr:$src))]>; def MOVZX64rm16 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "llgh\t{$dst, $src}", [(set GR64:$dst, (zextloadi64i16 rriaddr:$src))]>; def MOVZX64rm32 : Pseudo<(outs GR64:$dst), (ins rriaddr:$src), "llgf\t{$dst, $src}", [(set GR64:$dst, (zextloadi64i32 rriaddr:$src))]>; // truncstores def MOV32m8r : Pseudo<(outs), (ins rriaddr12:$dst, GR32:$src), "stc\t{$src, $dst}", [(truncstorei8 GR32:$src, rriaddr12:$dst)]>; def MOV32m8ry : Pseudo<(outs), (ins rriaddr:$dst, GR32:$src), "stcy\t{$src, $dst}", [(truncstorei8 GR32:$src, rriaddr:$dst)]>; def MOV32m16r : Pseudo<(outs), (ins rriaddr12:$dst, GR32:$src), "sth\t{$src, $dst}", [(truncstorei16 GR32:$src, rriaddr12:$dst)]>; def MOV32m16ry : Pseudo<(outs), (ins rriaddr:$dst, GR32:$src), "sthy\t{$src, $dst}", [(truncstorei16 GR32:$src, rriaddr:$dst)]>; def MOV64m8r : Pseudo<(outs), (ins rriaddr12:$dst, GR64:$src), "stc\t{$src, $dst}", [(truncstorei8 GR64:$src, rriaddr12:$dst)]>; def MOV64m8ry : Pseudo<(outs), (ins rriaddr:$dst, GR64:$src), "stcy\t{$src, $dst}", [(truncstorei8 GR64:$src, rriaddr:$dst)]>; def MOV64m16r : Pseudo<(outs), (ins rriaddr12:$dst, GR64:$src), "sth\t{$src, $dst}", [(truncstorei16 GR64:$src, rriaddr12:$dst)]>; def MOV64m16ry : Pseudo<(outs), (ins rriaddr:$dst, GR64:$src), "sthy\t{$src, $dst}", [(truncstorei16 GR64:$src, rriaddr:$dst)]>; def MOV64m32r : Pseudo<(outs), (ins rriaddr12:$dst, GR64:$src), "st\t{$src, $dst}", [(truncstorei32 GR64:$src, rriaddr12:$dst)]>; def MOV64m32ry : Pseudo<(outs), (ins rriaddr:$dst, GR64:$src), "sty\t{$src, $dst}", [(truncstorei32 GR64:$src, rriaddr:$dst)]>; // multiple regs moves // FIXME: should we use multiple arg nodes? def MOV32mrm : Pseudo<(outs), (ins riaddr:$dst, GR32:$from, GR32:$to), "stmy\t{$from, $to, $dst}", []>; def MOV64mrm : Pseudo<(outs), (ins riaddr:$dst, GR64:$from, GR64:$to), "stmg\t{$from, $to, $dst}", []>; def MOV32rmm : Pseudo<(outs GR32:$from, GR32:$to), (ins riaddr:$dst), "lmy\t{$from, $to, $dst}", []>; def MOV64rmm : Pseudo<(outs GR64:$from, GR64:$to), (ins riaddr:$dst), "lmg\t{$from, $to, $dst}", []>; let isReMaterializable = 1, isAsCheapAsAMove = 1, isTwoAddress = 1 in { def MOV64Pr0_even : Pseudo<(outs GR64P:$dst), (ins GR64P:$src), "lhi\t${dst:subreg_even}, 0", []>; def MOV128r0_even : Pseudo<(outs GR128:$dst), (ins GR128:$src), "lghi\t${dst:subreg_even}, 0", []>; } //===----------------------------------------------------------------------===// // Arithmetic Instructions let Defs = [PSW] in { def NEG32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src), "lcr\t{$dst, $src}", [(set GR32:$dst, (ineg GR32:$src)), (implicit PSW)]>; def NEG64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src), "lcgr\t{$dst, $src}", [(set GR64:$dst, (ineg GR64:$src)), (implicit PSW)]>; def NEG64rr32 : Pseudo<(outs GR64:$dst), (ins GR32:$src), "lcgfr\t{$dst, $src}", [(set GR64:$dst, (ineg (sext GR32:$src))), (implicit PSW)]>; } let isTwoAddress = 1 in { let Defs = [PSW] in { let isCommutable = 1 in { // X = ADD Y, Z == X = ADD Z, Y // FIXME: Provide proper encoding! def ADD32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "ar\t{$dst, $src2}", [(set GR32:$dst, (add GR32:$src1, GR32:$src2)), (implicit PSW)]>; def ADD64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "agr\t{$dst, $src2}", [(set GR64:$dst, (add GR64:$src1, GR64:$src2)), (implicit PSW)]>; } // FIXME: Provide proper encoding! def ADD32ri16 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, s16imm:$src2), "ahi\t{$dst, $src2}", [(set GR32:$dst, (add GR32:$src1, immSExt16:$src2)), (implicit PSW)]>; def ADD32ri : Pseudo<(outs GR32:$dst), (ins GR32:$src1, s32imm:$src2), "afi\t{$dst, $src2}", [(set GR32:$dst, (add GR32:$src1, imm:$src2)), (implicit PSW)]>; def ADD64ri16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, s16imm64:$src2), "aghi\t{$dst, $src2}", [(set GR64:$dst, (add GR64:$src1, immSExt16:$src2)), (implicit PSW)]>; def ADD64ri32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, s32imm64:$src2), "agfi\t{$dst, $src2}", [(set GR64:$dst, (add GR64:$src1, immSExt32:$src2)), (implicit PSW)]>; let isCommutable = 1 in { // X = AND Y, Z == X = AND Z, Y // FIXME: Provide proper encoding! def AND32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "nr\t{$dst, $src2}", [(set GR32:$dst, (and GR32:$src1, GR32:$src2))]>; def AND64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "ngr\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, GR64:$src2))]>; } // FIXME: Provide proper encoding! // FIXME: Compute masked bits properly! def AND32rill16 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "nill\t{$dst, $src2}", [(set GR32:$dst, (and GR32:$src1, i32ll16c:$src2))]>; def AND64rill16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nill\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64ll16c:$src2))]>; def AND32rilh16 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "nilh\t{$dst, $src2}", [(set GR32:$dst, (and GR32:$src1, i32lh16c:$src2))]>; def AND64rilh16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nilh\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64lh16c:$src2))]>; def AND64rihl16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nihl\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64hl16c:$src2))]>; def AND64rihh16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nihh\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64hh16c:$src2))]>; def AND32ri : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "nilf\t{$dst, $src2}", [(set GR32:$dst, (and GR32:$src1, imm:$src2))]>; def AND64rilo32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nilf\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64lo32c:$src2))]>; def AND64rihi32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "nihf\t{$dst, $src2}", [(set GR64:$dst, (and GR64:$src1, i64hi32c:$src2))]>; let isCommutable = 1 in { // X = OR Y, Z == X = OR Z, Y // FIXME: Provide proper encoding! def OR32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "or\t{$dst, $src2}", [(set GR32:$dst, (or GR32:$src1, GR32:$src2))]>; def OR64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "ogr\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, GR64:$src2))]>; } def OR32ri16 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "oill\t{$dst, $src2}", [(set GR32:$dst, (or GR32:$src1, i32ll16:$src2))]>; def OR32ri16h : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "oilh\t{$dst, $src2}", [(set GR32:$dst, (or GR32:$src1, i32lh16:$src2))]>; def OR32ri : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "oilf\t{$dst, $src2}", [(set GR32:$dst, (or GR32:$src1, imm:$src2))]>; def OR64rill16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oill\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64ll16:$src2))]>; def OR64rilh16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oilh\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64lh16:$src2))]>; def OR64rihl16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oihl\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64hl16:$src2))]>; def OR64rihh16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oihh\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64hh16:$src2))]>; def OR64rilo32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oilf\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64lo32:$src2))]>; def OR64rihi32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, i64imm:$src2), "oihf\t{$dst, $src2}", [(set GR64:$dst, (or GR64:$src1, i64hi32:$src2))]>; // FIXME: Provide proper encoding! def SUB32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "sr\t{$dst, $src2}", [(set GR32:$dst, (sub GR32:$src1, GR32:$src2))]>; def SUB64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "sgr\t{$dst, $src2}", [(set GR64:$dst, (sub GR64:$src1, GR64:$src2))]>; let isCommutable = 1 in { // X = XOR Y, Z == X = XOR Z, Y // FIXME: Provide proper encoding! def XOR32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "xr\t{$dst, $src2}", [(set GR32:$dst, (xor GR32:$src1, GR32:$src2))]>; def XOR64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "xgr\t{$dst, $src2}", [(set GR64:$dst, (xor GR64:$src1, GR64:$src2))]>; } def XOR32ri : Pseudo<(outs GR32:$dst), (ins GR32:$src1, i32imm:$src2), "xilf\t{$dst, $src2}", [(set GR32:$dst, (xor GR32:$src1, imm:$src2))]>; } // Defs = [PSW] let isCommutable = 1 in { // X = MUL Y, Z == X = MUL Z, Y def MUL32rr : Pseudo<(outs GR32:$dst), (ins GR32:$src1, GR32:$src2), "msr\t{$dst, $src2}", [(set GR32:$dst, (mul GR32:$src1, GR32:$src2))]>; def MUL64rr : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR64:$src2), "msgr\t{$dst, $src2}", [(set GR64:$dst, (mul GR64:$src1, GR64:$src2))]>; } def MUL64rrP : Pseudo<(outs GR64P:$dst), (ins GR64P:$src1, GR32:$src2), "mr\t{$dst, $src2}", []>; def UMUL64rrP : Pseudo<(outs GR64P:$dst), (ins GR64P:$src1, GR32:$src2), "mlr\t{$dst, $src2}", []>; def UMUL128rrP : Pseudo<(outs GR128:$dst), (ins GR128:$src1, GR64:$src2), "mlgr\t{$dst, $src2}", []>; def MUL32ri16 : Pseudo<(outs GR32:$dst), (ins GR32:$src1, s16imm:$src2), "mhi\t{$dst, $src2}", [(set GR32:$dst, (mul GR32:$src1, i32immSExt16:$src2))]>; def MUL64ri16 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, s16imm64:$src2), "mghi\t{$dst, $src2}", [(set GR64:$dst, (mul GR64:$src1, immSExt16:$src2))]>; def MUL32ri : Pseudo<(outs GR32:$dst), (ins GR32:$src1, s32imm:$src2), "msfi\t{$dst, $src2}", [(set GR32:$dst, (mul GR32:$src1, imm:$src2))]>, Requires<[IsZ10]>; def MUL64ri32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, s32imm64:$src2), "msgfi\t{$dst, $src2}", [(set GR64:$dst, (mul GR64:$src1, i64immSExt32:$src2))]>, Requires<[IsZ10]>; def MUL32rm : Pseudo<(outs GR32:$dst), (ins GR32:$src1, rriaddr12:$src2), "ms\t{$dst, $src2}", [(set GR32:$dst, (mul GR32:$src1, (load rriaddr12:$src2)))]>; def MUL32rmy : Pseudo<(outs GR32:$dst), (ins GR32:$src1, rriaddr:$src2), "msy\t{$dst, $src2}", [(set GR32:$dst, (mul GR32:$src1, (load rriaddr:$src2)))]>; def MUL64rm : Pseudo<(outs GR64:$dst), (ins GR64:$src1, rriaddr:$src2), "msg\t{$dst, $src2}", [(set GR64:$dst, (mul GR64:$src1, (load rriaddr:$src2)))]>; def MULSX64rr32 : Pseudo<(outs GR64:$dst), (ins GR64:$src1, GR32:$src2), "msgfr\t{$dst, $src2}", [(set GR64:$dst, (mul GR64:$src1, (sext GR32:$src2)))]>; def SDIVREM32r : Pseudo<(outs GR128:$dst), (ins GR128:$src1, GR32:$src2), "dsgfr\t{$dst, $src2}", []>; def SDIVREM64r : Pseudo<(outs GR128:$dst), (ins GR128:$src1, GR64:$src2), "dsgr\t{$dst, $src2}", []>; def UDIVREM32r : Pseudo<(outs GR64P:$dst), (ins GR64P:$src1, GR32:$src2), "dlr\t{$dst, $src2}", []>; def UDIVREM64r : Pseudo<(outs GR128:$dst), (ins GR128:$src1, GR64:$src2), "dlgr\t{$dst, $src2}", []>; let mayLoad = 1 in { def SDIVREM32m : Pseudo<(outs GR128:$dst), (ins GR128:$src1, rriaddr:$src2), "dsgf\t{$dst, $src2}", []>; def SDIVREM64m : Pseudo<(outs GR128:$dst), (ins GR128:$src1, rriaddr:$src2), "dsg\t{$dst, $src2}", []>; def UDIVREM32m : Pseudo<(outs GR64P:$dst), (ins GR64P:$src1, rriaddr:$src2), "dl\t{$dst, $src2}", []>; def UDIVREM64m : Pseudo<(outs GR128:$dst), (ins GR128:$src1, rriaddr:$src2), "dlg\t{$dst, $src2}", []>; } // mayLoad } // isTwoAddress = 1 //===----------------------------------------------------------------------===// // Shifts let isTwoAddress = 1 in def SRL32rri : Pseudo<(outs GR32:$dst), (ins GR32:$src, riaddr32:$amt), "srl\t{$src, $amt}", [(set GR32:$dst, (srl GR32:$src, riaddr32:$amt))]>; def SRL64rri : Pseudo<(outs GR64:$dst), (ins GR64:$src, riaddr:$amt), "srlg\t{$dst, $src, $amt}", [(set GR64:$dst, (srl GR64:$src, riaddr:$amt))]>; let isTwoAddress = 1 in def SHL32rri : Pseudo<(outs GR32:$dst), (ins GR32:$src, riaddr32:$amt), "sll\t{$src, $amt}", [(set GR32:$dst, (shl GR32:$src, riaddr32:$amt))]>; def SHL64rri : Pseudo<(outs GR64:$dst), (ins GR64:$src, riaddr:$amt), "sllg\t{$dst, $src, $amt}", [(set GR64:$dst, (shl GR64:$src, riaddr:$amt))]>; let Defs = [PSW] in { let isTwoAddress = 1 in def SRA32rri : Pseudo<(outs GR32:$dst), (ins GR32:$src, riaddr32:$amt), "sra\t{$src, $amt}", [(set GR32:$dst, (sra GR32:$src, riaddr32:$amt)), (implicit PSW)]>; def SRA64rri : Pseudo<(outs GR64:$dst), (ins GR64:$src, riaddr:$amt), "srag\t{$dst, $src, $amt}", [(set GR64:$dst, (sra GR64:$src, riaddr:$amt)), (implicit PSW)]>; } // Defs = [PSW] def ROTL32rri : Pseudo<(outs GR32:$dst), (ins GR32:$src, riaddr32:$amt), "rll\t{$dst, $src, $amt}", [(set GR32:$dst, (rotl GR32:$src, riaddr32:$amt))]>; def ROTL64rri : Pseudo<(outs GR64:$dst), (ins GR64:$src, riaddr:$amt), "rllg\t{$dst, $src, $amt}", [(set GR64:$dst, (rotl GR64:$src, riaddr:$amt))]>; //===----------------------------------------------------------------------===// // Test instructions (like AND but do not produce any result) // Integer comparisons let Defs = [PSW] in { def CMP32rr : Pseudo<(outs), (ins GR32:$src1, GR32:$src2), "cr\t$src1, $src2", [(SystemZcmp GR32:$src1, GR32:$src2), (implicit PSW)]>; def CMP64rr : Pseudo<(outs), (ins GR64:$src1, GR64:$src2), "cgr\t$src1, $src2", [(SystemZcmp GR64:$src1, GR64:$src2), (implicit PSW)]>; def CMP32ri : Pseudo<(outs), (ins GR32:$src1, s32imm:$src2), "cfi\t$src1, $src2", [(SystemZcmp GR32:$src1, imm:$src2), (implicit PSW)]>; def CMP64ri32 : Pseudo<(outs), (ins GR64:$src1, s32imm64:$src2), "cgfi\t$src1, $src2", [(SystemZcmp GR64:$src1, i64immSExt32:$src2), (implicit PSW)]>; def CMP32rm : Pseudo<(outs), (ins GR32:$src1, rriaddr12:$src2), "c\t$src1, $src2", [(SystemZcmp GR32:$src1, (load rriaddr12:$src2)), (implicit PSW)]>; def CMP32rmy : Pseudo<(outs), (ins GR32:$src1, rriaddr:$src2), "cy\t$src1, $src2", [(SystemZcmp GR32:$src1, (load rriaddr:$src2)), (implicit PSW)]>; def CMP64rm : Pseudo<(outs), (ins GR64:$src1, rriaddr:$src2), "cg\t$src1, $src2", [(SystemZcmp GR64:$src1, (load rriaddr:$src2)), (implicit PSW)]>; def UCMP32rr : Pseudo<(outs), (ins GR32:$src1, GR32:$src2), "clr\t$src1, $src2", [(SystemZucmp GR32:$src1, GR32:$src2), (implicit PSW)]>; def UCMP64rr : Pseudo<(outs), (ins GR64:$src1, GR64:$src2), "clgr\t$src1, $src2", [(SystemZucmp GR64:$src1, GR64:$src2), (implicit PSW)]>; def UCMP32ri : Pseudo<(outs), (ins GR32:$src1, i32imm:$src2), "clfi\t$src1, $src2", [(SystemZucmp GR32:$src1, imm:$src2), (implicit PSW)]>; def UCMP64ri32 : Pseudo<(outs), (ins GR64:$src1, i64i32imm:$src2), "clgfi\t$src1, $src2", [(SystemZucmp GR64:$src1, i64immZExt32:$src2), (implicit PSW)]>; def UCMP32rm : Pseudo<(outs), (ins GR32:$src1, rriaddr12:$src2), "cl\t$src1, $src2", [(SystemZucmp GR32:$src1, (load rriaddr12:$src2)), (implicit PSW)]>; def UCMP32rmy : Pseudo<(outs), (ins GR32:$src1, rriaddr:$src2), "cly\t$src1, $src2", [(SystemZucmp GR32:$src1, (load rriaddr:$src2)), (implicit PSW)]>; def UCMP64rm : Pseudo<(outs), (ins GR64:$src1, rriaddr:$src2), "clg\t$src1, $src2", [(SystemZucmp GR64:$src1, (load rriaddr:$src2)), (implicit PSW)]>; def CMPSX64rr32 : Pseudo<(outs), (ins GR64:$src1, GR32:$src2), "cgfr\t$src1, $src2", [(SystemZucmp GR64:$src1, (sext GR32:$src2)), (implicit PSW)]>; def UCMPZX64rr32 : Pseudo<(outs), (ins GR64:$src1, GR32:$src2), "clgfr\t$src1, $src2", [(SystemZucmp GR64:$src1, (zext GR32:$src2)), (implicit PSW)]>; def CMPSX64rm32 : Pseudo<(outs), (ins GR64:$src1, rriaddr:$src2), "cgf\t$src1, $src2", [(SystemZucmp GR64:$src1, (sextloadi64i32 rriaddr:$src2)), (implicit PSW)]>; def UCMPZX64rm32 : Pseudo<(outs), (ins GR64:$src1, rriaddr:$src2), "clgf\t$src1, $src2", [(SystemZucmp GR64:$src1, (zextloadi64i32 rriaddr:$src2)), (implicit PSW)]>; // FIXME: Add other crazy ucmp forms } // Defs = [PSW] //===----------------------------------------------------------------------===// // Non-Instruction Patterns. //===----------------------------------------------------------------------===// // ConstPools, JumpTables def : Pat<(SystemZpcrelwrapper tjumptable:$src), (LA64rm tjumptable:$src)>; def : Pat<(SystemZpcrelwrapper tconstpool:$src), (LA64rm tconstpool:$src)>; // anyext def : Pat<(i64 (anyext GR32:$src)), (MOVZX64rr32 GR32:$src)>; // calls def : Pat<(SystemZcall (i64 tglobaladdr:$dst)), (CALLi tglobaladdr:$dst)>; def : Pat<(SystemZcall (i64 texternalsym:$dst)), (CALLi texternalsym:$dst)>; //===----------------------------------------------------------------------===// // Peepholes. //===----------------------------------------------------------------------===// // FIXME: use add/sub tricks with 32678/-32768 // Arbitrary immediate support. Implement in terms of LLIHF/OILF. def : Pat<(i64 imm:$imm), (OR64rilo32 (MOV64rihi32 (HI32 imm:$imm)), (LO32 imm:$imm))>; // trunc patterns def : Pat<(i32 (trunc GR64:$src)), (EXTRACT_SUBREG GR64:$src, subreg_32bit)>; // sext_inreg patterns def : Pat<(sext_inreg GR64:$src, i32), (MOVSX64rr32 (EXTRACT_SUBREG GR64:$src, subreg_32bit))>; // extload patterns def : Pat<(extloadi32i8 rriaddr:$src), (MOVZX32rm8 rriaddr:$src)>; def : Pat<(extloadi32i16 rriaddr:$src), (MOVZX32rm16 rriaddr:$src)>; def : Pat<(extloadi64i8 rriaddr:$src), (MOVZX64rm8 rriaddr:$src)>; def : Pat<(extloadi64i16 rriaddr:$src), (MOVZX64rm16 rriaddr:$src)>; def : Pat<(extloadi64i32 rriaddr:$src), (MOVZX64rm32 rriaddr:$src)>; // muls def : Pat<(mulhs GR32:$src1, GR32:$src2), (EXTRACT_SUBREG (MUL64rrP (INSERT_SUBREG (i64 (IMPLICIT_DEF)), GR32:$src1, subreg_odd32), GR32:$src2), subreg_even32)>; def : Pat<(mulhu GR32:$src1, GR32:$src2), (EXTRACT_SUBREG (UMUL64rrP (INSERT_SUBREG (i64 (IMPLICIT_DEF)), GR32:$src1, subreg_odd32), GR32:$src2), subreg_even32)>; def : Pat<(mulhu GR64:$src1, GR64:$src2), (EXTRACT_SUBREG (UMUL128rrP (INSERT_SUBREG (i128 (IMPLICIT_DEF)), GR64:$src1, subreg_odd), GR64:$src2), subreg_even)>; def : Pat<(i32 imm:$src), (EXTRACT_SUBREG (MOV64ri32 (i64 imm:$src)), subreg_32bit)>;