| //===-- BUFInstructions.td - Buffer Instruction Defintions ----------------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| def MUBUFAddr32 : ComplexPattern<i64, 9, "SelectMUBUFAddr32">; |
| def MUBUFAddr64 : ComplexPattern<i64, 7, "SelectMUBUFAddr64">; |
| def MUBUFAddr64Atomic : ComplexPattern<i64, 5, "SelectMUBUFAddr64">; |
| |
| def MUBUFScratchOffen : ComplexPattern<i64, 4, "SelectMUBUFScratchOffen", [], [SDNPWantParent]>; |
| def MUBUFScratchOffset : ComplexPattern<i64, 3, "SelectMUBUFScratchOffset", [], [SDNPWantParent], 20>; |
| |
| def MUBUFOffset : ComplexPattern<i64, 6, "SelectMUBUFOffset">; |
| def MUBUFOffsetNoGLC : ComplexPattern<i64, 3, "SelectMUBUFOffset">; |
| def MUBUFOffsetAtomic : ComplexPattern<i64, 4, "SelectMUBUFOffset">; |
| def MUBUFIntrinsicOffset : ComplexPattern<i32, 2, "SelectMUBUFIntrinsicOffset">; |
| def MUBUFIntrinsicVOffset : ComplexPattern<i32, 3, "SelectMUBUFIntrinsicVOffset">; |
| |
| class MubufLoad <SDPatternOperator op> : PatFrag < |
| (ops node:$ptr), (op node:$ptr), [{ |
| auto const AS = cast<MemSDNode>(N)->getAddressSpace(); |
| return AS == AMDGPUASI.GLOBAL_ADDRESS || |
| AS == AMDGPUASI.CONSTANT_ADDRESS; |
| }]>; |
| |
| def mubuf_load : MubufLoad <load>; |
| def mubuf_az_extloadi8 : MubufLoad <az_extloadi8>; |
| def mubuf_sextloadi8 : MubufLoad <sextloadi8>; |
| def mubuf_az_extloadi16 : MubufLoad <az_extloadi16>; |
| def mubuf_sextloadi16 : MubufLoad <sextloadi16>; |
| def mubuf_load_atomic : MubufLoad <atomic_load>; |
| |
| def BUFAddrKind { |
| int Offset = 0; |
| int OffEn = 1; |
| int IdxEn = 2; |
| int BothEn = 3; |
| int Addr64 = 4; |
| } |
| |
| class getAddrName<int addrKind> { |
| string ret = |
| !if(!eq(addrKind, BUFAddrKind.Offset), "offset", |
| !if(!eq(addrKind, BUFAddrKind.OffEn), "offen", |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), "idxen", |
| !if(!eq(addrKind, BUFAddrKind.BothEn), "bothen", |
| !if(!eq(addrKind, BUFAddrKind.Addr64), "addr64", |
| ""))))); |
| } |
| |
| class MUBUFAddr64Table <bit is_addr64, string Name> { |
| bit IsAddr64 = is_addr64; |
| string OpName = Name; |
| } |
| |
| class MUBUFLdsTable <bit is_lds, string Name> { |
| bit IsLds = is_lds; |
| string OpName = Name; |
| } |
| |
| class MTBUFAddr64Table <bit is_addr64, string Name> { |
| bit IsAddr64 = is_addr64; |
| string OpName = Name; |
| } |
| |
| //===----------------------------------------------------------------------===// |
| // MTBUF classes |
| //===----------------------------------------------------------------------===// |
| |
| class MTBUF_Pseudo <string opName, dag outs, dag ins, |
| string asmOps, list<dag> pattern=[]> : |
| InstSI<outs, ins, "", pattern>, |
| SIMCInstr<opName, SIEncodingFamily.NONE> { |
| |
| let isPseudo = 1; |
| let isCodeGenOnly = 1; |
| let Size = 8; |
| let UseNamedOperandTable = 1; |
| |
| string Mnemonic = opName; |
| string AsmOperands = asmOps; |
| |
| let VM_CNT = 1; |
| let EXP_CNT = 1; |
| let MTBUF = 1; |
| let Uses = [EXEC]; |
| let hasSideEffects = 0; |
| let SchedRW = [WriteVMEM]; |
| |
| let AsmMatchConverter = "cvtMtbuf"; |
| |
| bits<1> offen = 0; |
| bits<1> idxen = 0; |
| bits<1> addr64 = 0; |
| bits<1> has_vdata = 1; |
| bits<1> has_vaddr = 1; |
| bits<1> has_glc = 1; |
| bits<1> glc_value = 0; // the value for glc if no such operand |
| bits<4> dfmt_value = 1; // the value for dfmt if no such operand |
| bits<3> nfmt_value = 0; // the value for nfmt if no such operand |
| bits<1> has_srsrc = 1; |
| bits<1> has_soffset = 1; |
| bits<1> has_offset = 1; |
| bits<1> has_slc = 1; |
| bits<1> has_tfe = 1; |
| bits<1> has_dfmt = 1; |
| bits<1> has_nfmt = 1; |
| } |
| |
| class MTBUF_Real <MTBUF_Pseudo ps> : |
| InstSI <ps.OutOperandList, ps.InOperandList, ps.Mnemonic # ps.AsmOperands, []> { |
| |
| let isPseudo = 0; |
| let isCodeGenOnly = 0; |
| |
| // copy relevant pseudo op flags |
| let SubtargetPredicate = ps.SubtargetPredicate; |
| let AsmMatchConverter = ps.AsmMatchConverter; |
| let Constraints = ps.Constraints; |
| let DisableEncoding = ps.DisableEncoding; |
| let TSFlags = ps.TSFlags; |
| |
| bits<12> offset; |
| bits<1> glc; |
| bits<4> dfmt; |
| bits<3> nfmt; |
| bits<8> vaddr; |
| bits<8> vdata; |
| bits<7> srsrc; |
| bits<1> slc; |
| bits<1> tfe; |
| bits<8> soffset; |
| } |
| |
| class getMTBUFInsDA<list<RegisterClass> vdataList, |
| list<RegisterClass> vaddrList=[]> { |
| RegisterClass vdataClass = !if(!empty(vdataList), ?, !head(vdataList)); |
| RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); |
| dag InsNoData = !if(!empty(vaddrList), |
| (ins SReg_128:$srsrc, SCSrc_b32:$soffset, |
| offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe), |
| (ins vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, |
| offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe) |
| ); |
| dag InsData = !if(!empty(vaddrList), |
| (ins vdataClass:$vdata, SReg_128:$srsrc, |
| SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, |
| SLC:$slc, TFE:$tfe), |
| (ins vdataClass:$vdata, vaddrClass:$vaddr, SReg_128:$srsrc, |
| SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, |
| SLC:$slc, TFE:$tfe) |
| ); |
| dag ret = !if(!empty(vdataList), InsNoData, InsData); |
| } |
| |
| class getMTBUFIns<int addrKind, list<RegisterClass> vdataList=[]> { |
| dag ret = |
| !if(!eq(addrKind, BUFAddrKind.Offset), getMTBUFInsDA<vdataList>.ret, |
| !if(!eq(addrKind, BUFAddrKind.OffEn), getMTBUFInsDA<vdataList, [VGPR_32]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), getMTBUFInsDA<vdataList, [VGPR_32]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), getMTBUFInsDA<vdataList, [VReg_64]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.Addr64), getMTBUFInsDA<vdataList, [VReg_64]>.ret, |
| (ins)))))); |
| } |
| |
| class getMTBUFAsmOps<int addrKind> { |
| string Pfx = |
| !if(!eq(addrKind, BUFAddrKind.Offset), "off, $srsrc, $dfmt, $nfmt, $soffset", |
| !if(!eq(addrKind, BUFAddrKind.OffEn), |
| "$vaddr, $srsrc, $dfmt, $nfmt, $soffset offen", |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), |
| "$vaddr, $srsrc, $dfmt, $nfmt, $soffset idxen", |
| !if(!eq(addrKind, BUFAddrKind.BothEn), |
| "$vaddr, $srsrc, $dfmt, $nfmt, $soffset idxen offen", |
| !if(!eq(addrKind, BUFAddrKind.Addr64), |
| "$vaddr, $srsrc, $dfmt, $nfmt, $soffset addr64", |
| ""))))); |
| string ret = Pfx # "$offset"; |
| } |
| |
| class MTBUF_SetupAddr<int addrKind> { |
| bits<1> offen = !if(!eq(addrKind, BUFAddrKind.OffEn), 1, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); |
| |
| bits<1> idxen = !if(!eq(addrKind, BUFAddrKind.IdxEn), 1, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); |
| |
| bits<1> addr64 = !if(!eq(addrKind, BUFAddrKind.Addr64), 1, 0); |
| |
| bits<1> has_vaddr = !if(!eq(addrKind, BUFAddrKind.Offset), 0, 1); |
| } |
| |
| class MTBUF_Load_Pseudo <string opName, |
| int addrKind, |
| RegisterClass vdataClass, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind> |
| : MTBUF_Pseudo<opName, |
| (outs vdataClass:$vdata), |
| getMTBUFIns<addrKindCopy>.ret, |
| " $vdata, " # getMTBUFAsmOps<addrKindCopy>.ret # "$glc$slc$tfe", |
| pattern>, |
| MTBUF_SetupAddr<addrKindCopy> { |
| let PseudoInstr = opName # "_" # getAddrName<addrKindCopy>.ret; |
| let mayLoad = 1; |
| let mayStore = 0; |
| } |
| |
| multiclass MTBUF_Pseudo_Loads<string opName, RegisterClass vdataClass, |
| ValueType load_vt = i32, |
| SDPatternOperator ld = null_frag> { |
| |
| def _OFFSET : MTBUF_Load_Pseudo <opName, BUFAddrKind.Offset, vdataClass, |
| [(set load_vt:$vdata, |
| (ld (MUBUFOffset v4i32:$srsrc, i32:$soffset, i16:$offset, i8:$dfmt, |
| i8:$nfmt, i1:$glc, i1:$slc, i1:$tfe)))]>, |
| MTBUFAddr64Table<0, NAME>; |
| |
| def _ADDR64 : MTBUF_Load_Pseudo <opName, BUFAddrKind.Addr64, vdataClass, |
| [(set load_vt:$vdata, |
| (ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, |
| i8:$dfmt, i8:$nfmt, i1:$glc, i1:$slc, i1:$tfe)))]>, |
| MTBUFAddr64Table<1, NAME>; |
| |
| def _OFFEN : MTBUF_Load_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN : MTBUF_Load_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN : MTBUF_Load_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| |
| let DisableWQM = 1 in { |
| def _OFFSET_exact : MTBUF_Load_Pseudo <opName, BUFAddrKind.Offset, vdataClass>; |
| def _OFFEN_exact : MTBUF_Load_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN_exact : MTBUF_Load_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN_exact : MTBUF_Load_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| } |
| } |
| |
| class MTBUF_Store_Pseudo <string opName, |
| int addrKind, |
| RegisterClass vdataClass, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind, |
| RegisterClass vdataClassCopy = vdataClass> |
| : MTBUF_Pseudo<opName, |
| (outs), |
| getMTBUFIns<addrKindCopy, [vdataClassCopy]>.ret, |
| " $vdata, " # getMTBUFAsmOps<addrKindCopy>.ret # "$glc$slc$tfe", |
| pattern>, |
| MTBUF_SetupAddr<addrKindCopy> { |
| let PseudoInstr = opName # "_" # getAddrName<addrKindCopy>.ret; |
| let mayLoad = 0; |
| let mayStore = 1; |
| } |
| |
| multiclass MTBUF_Pseudo_Stores<string opName, RegisterClass vdataClass, |
| ValueType store_vt = i32, |
| SDPatternOperator st = null_frag> { |
| |
| def _OFFSET : MTBUF_Store_Pseudo <opName, BUFAddrKind.Offset, vdataClass, |
| [(st store_vt:$vdata, (MUBUFOffset v4i32:$srsrc, i32:$soffset, |
| i16:$offset, i8:$dfmt, i8:$nfmt, i1:$glc, |
| i1:$slc, i1:$tfe))]>, |
| MTBUFAddr64Table<0, NAME>; |
| |
| def _ADDR64 : MTBUF_Store_Pseudo <opName, BUFAddrKind.Addr64, vdataClass, |
| [(st store_vt:$vdata, (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, |
| i16:$offset, i8:$dfmt, i8:$nfmt, i1:$glc, |
| i1:$slc, i1:$tfe))]>, |
| MTBUFAddr64Table<1, NAME>; |
| |
| def _OFFEN : MTBUF_Store_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN : MTBUF_Store_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN : MTBUF_Store_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| |
| let DisableWQM = 1 in { |
| def _OFFSET_exact : MTBUF_Store_Pseudo <opName, BUFAddrKind.Offset, vdataClass>; |
| def _OFFEN_exact : MTBUF_Store_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN_exact : MTBUF_Store_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN_exact : MTBUF_Store_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| } |
| } |
| |
| |
| //===----------------------------------------------------------------------===// |
| // MUBUF classes |
| //===----------------------------------------------------------------------===// |
| |
| class MUBUF_Pseudo <string opName, dag outs, dag ins, |
| string asmOps, list<dag> pattern=[]> : |
| InstSI<outs, ins, "", pattern>, |
| SIMCInstr<opName, SIEncodingFamily.NONE> { |
| |
| let isPseudo = 1; |
| let isCodeGenOnly = 1; |
| let Size = 8; |
| let UseNamedOperandTable = 1; |
| |
| string Mnemonic = opName; |
| string AsmOperands = asmOps; |
| |
| let VM_CNT = 1; |
| let EXP_CNT = 1; |
| let MUBUF = 1; |
| let Uses = [EXEC]; |
| let hasSideEffects = 0; |
| let SchedRW = [WriteVMEM]; |
| |
| let AsmMatchConverter = "cvtMubuf"; |
| |
| bits<1> offen = 0; |
| bits<1> idxen = 0; |
| bits<1> addr64 = 0; |
| bits<1> lds = 0; |
| bits<1> has_vdata = 1; |
| bits<1> has_vaddr = 1; |
| bits<1> has_glc = 1; |
| bits<1> glc_value = 0; // the value for glc if no such operand |
| bits<1> has_srsrc = 1; |
| bits<1> has_soffset = 1; |
| bits<1> has_offset = 1; |
| bits<1> has_slc = 1; |
| bits<1> has_tfe = 1; |
| } |
| |
| class MUBUF_Real <bits<7> op, MUBUF_Pseudo ps> : |
| InstSI <ps.OutOperandList, ps.InOperandList, ps.Mnemonic # ps.AsmOperands, []> { |
| |
| let isPseudo = 0; |
| let isCodeGenOnly = 0; |
| |
| // copy relevant pseudo op flags |
| let SubtargetPredicate = ps.SubtargetPredicate; |
| let AsmMatchConverter = ps.AsmMatchConverter; |
| let Constraints = ps.Constraints; |
| let DisableEncoding = ps.DisableEncoding; |
| let TSFlags = ps.TSFlags; |
| |
| bits<12> offset; |
| bits<1> glc; |
| bits<8> vaddr; |
| bits<8> vdata; |
| bits<7> srsrc; |
| bits<1> slc; |
| bits<1> tfe; |
| bits<8> soffset; |
| } |
| |
| |
| // For cache invalidation instructions. |
| class MUBUF_Invalidate <string opName, SDPatternOperator node> : |
| MUBUF_Pseudo<opName, (outs), (ins), "", [(node)]> { |
| |
| let AsmMatchConverter = ""; |
| |
| let hasSideEffects = 1; |
| let mayStore = 1; |
| |
| // Set everything to 0. |
| let offen = 0; |
| let idxen = 0; |
| let addr64 = 0; |
| let has_vdata = 0; |
| let has_vaddr = 0; |
| let has_glc = 0; |
| let glc_value = 0; |
| let has_srsrc = 0; |
| let has_soffset = 0; |
| let has_offset = 0; |
| let has_slc = 0; |
| let has_tfe = 0; |
| } |
| |
| class getMUBUFInsDA<list<RegisterClass> vdataList, |
| list<RegisterClass> vaddrList=[], |
| bit isLds = 0> { |
| RegisterClass vdataClass = !if(!empty(vdataList), ?, !head(vdataList)); |
| RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); |
| dag InsNoData = !if(!empty(vaddrList), |
| (ins SReg_128:$srsrc, SCSrc_b32:$soffset, |
| offset:$offset, GLC:$glc, SLC:$slc), |
| (ins vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, |
| offset:$offset, GLC:$glc, SLC:$slc) |
| ); |
| dag InsData = !if(!empty(vaddrList), |
| (ins vdataClass:$vdata, SReg_128:$srsrc, |
| SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc), |
| (ins vdataClass:$vdata, vaddrClass:$vaddr, SReg_128:$srsrc, |
| SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc) |
| ); |
| dag ret = !con( |
| !if(!empty(vdataList), InsNoData, InsData), |
| !if(isLds, (ins), (ins TFE:$tfe)) |
| ); |
| } |
| |
| class getMUBUFIns<int addrKind, list<RegisterClass> vdataList=[], bit isLds = 0> { |
| dag ret = |
| !if(!eq(addrKind, BUFAddrKind.Offset), getMUBUFInsDA<vdataList, [], isLds>.ret, |
| !if(!eq(addrKind, BUFAddrKind.OffEn), getMUBUFInsDA<vdataList, [VGPR_32], isLds>.ret, |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), getMUBUFInsDA<vdataList, [VGPR_32], isLds>.ret, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), getMUBUFInsDA<vdataList, [VReg_64], isLds>.ret, |
| !if(!eq(addrKind, BUFAddrKind.Addr64), getMUBUFInsDA<vdataList, [VReg_64], isLds>.ret, |
| (ins)))))); |
| } |
| |
| class getMUBUFAsmOps<int addrKind> { |
| string Pfx = |
| !if(!eq(addrKind, BUFAddrKind.Offset), "off, $srsrc, $soffset", |
| !if(!eq(addrKind, BUFAddrKind.OffEn), "$vaddr, $srsrc, $soffset offen", |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), "$vaddr, $srsrc, $soffset idxen", |
| !if(!eq(addrKind, BUFAddrKind.BothEn), "$vaddr, $srsrc, $soffset idxen offen", |
| !if(!eq(addrKind, BUFAddrKind.Addr64), "$vaddr, $srsrc, $soffset addr64", |
| ""))))); |
| string ret = Pfx # "$offset"; |
| } |
| |
| class MUBUF_SetupAddr<int addrKind> { |
| bits<1> offen = !if(!eq(addrKind, BUFAddrKind.OffEn), 1, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); |
| |
| bits<1> idxen = !if(!eq(addrKind, BUFAddrKind.IdxEn), 1, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); |
| |
| bits<1> addr64 = !if(!eq(addrKind, BUFAddrKind.Addr64), 1, 0); |
| |
| bits<1> has_vaddr = !if(!eq(addrKind, BUFAddrKind.Offset), 0, 1); |
| } |
| |
| class MUBUF_Load_Pseudo <string opName, |
| int addrKind, |
| RegisterClass vdataClass, |
| bit HasTiedDest = 0, |
| bit isLds = 0, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind> |
| : MUBUF_Pseudo<opName, |
| (outs vdataClass:$vdata), |
| !con(getMUBUFIns<addrKindCopy, [], isLds>.ret, |
| !if(HasTiedDest, (ins vdataClass:$vdata_in), (ins))), |
| " $vdata, " # getMUBUFAsmOps<addrKindCopy>.ret # "$glc$slc" # |
| !if(isLds, " lds", "$tfe"), |
| pattern>, |
| MUBUF_SetupAddr<addrKindCopy> { |
| let PseudoInstr = opName # !if(isLds, "_lds", "") # |
| "_" # getAddrName<addrKindCopy>.ret; |
| let AsmMatchConverter = !if(isLds, "cvtMubufLds", "cvtMubuf"); |
| |
| let Constraints = !if(HasTiedDest, "$vdata = $vdata_in", ""); |
| let mayLoad = 1; |
| let mayStore = 0; |
| let maybeAtomic = 1; |
| let Uses = !if(isLds, [EXEC, M0], [EXEC]); |
| let has_tfe = !if(isLds, 0, 1); |
| let lds = isLds; |
| } |
| |
| // FIXME: tfe can't be an operand because it requires a separate |
| // opcode because it needs an N+1 register class dest register. |
| multiclass MUBUF_Pseudo_Loads<string opName, RegisterClass vdataClass, |
| ValueType load_vt = i32, |
| SDPatternOperator ld = null_frag, |
| bit TiedDest = 0, |
| bit isLds = 0> { |
| |
| def _OFFSET : MUBUF_Load_Pseudo <opName, BUFAddrKind.Offset, vdataClass, |
| TiedDest, isLds, |
| !if(isLds, |
| [], |
| [(set load_vt:$vdata, |
| (ld (MUBUFOffset v4i32:$srsrc, i32:$soffset, i16:$offset, i1:$glc, i1:$slc, i1:$tfe)))])>, |
| MUBUFAddr64Table<0, NAME # !if(isLds, "_LDS", "")>; |
| |
| def _ADDR64 : MUBUF_Load_Pseudo <opName, BUFAddrKind.Addr64, vdataClass, |
| TiedDest, isLds, |
| !if(isLds, |
| [], |
| [(set load_vt:$vdata, |
| (ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, i1:$glc, i1:$slc, i1:$tfe)))])>, |
| MUBUFAddr64Table<1, NAME # !if(isLds, "_LDS", "")>; |
| |
| def _OFFEN : MUBUF_Load_Pseudo <opName, BUFAddrKind.OffEn, vdataClass, TiedDest, isLds>; |
| def _IDXEN : MUBUF_Load_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass, TiedDest, isLds>; |
| def _BOTHEN : MUBUF_Load_Pseudo <opName, BUFAddrKind.BothEn, vdataClass, TiedDest, isLds>; |
| |
| let DisableWQM = 1 in { |
| def _OFFSET_exact : MUBUF_Load_Pseudo <opName, BUFAddrKind.Offset, vdataClass, TiedDest, isLds>; |
| def _OFFEN_exact : MUBUF_Load_Pseudo <opName, BUFAddrKind.OffEn, vdataClass, TiedDest, isLds>; |
| def _IDXEN_exact : MUBUF_Load_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass, TiedDest, isLds>; |
| def _BOTHEN_exact : MUBUF_Load_Pseudo <opName, BUFAddrKind.BothEn, vdataClass, TiedDest, isLds>; |
| } |
| } |
| |
| multiclass MUBUF_Pseudo_Loads_Lds<string opName, RegisterClass vdataClass, |
| ValueType load_vt = i32, |
| SDPatternOperator ld_nolds = null_frag, |
| SDPatternOperator ld_lds = null_frag> { |
| defm NAME : MUBUF_Pseudo_Loads<opName, vdataClass, load_vt, ld_nolds>; |
| defm _LDS : MUBUF_Pseudo_Loads<opName, vdataClass, load_vt, ld_lds, 0, 1>; |
| } |
| |
| class MUBUF_Store_Pseudo <string opName, |
| int addrKind, |
| RegisterClass vdataClass, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind, |
| RegisterClass vdataClassCopy = vdataClass> |
| : MUBUF_Pseudo<opName, |
| (outs), |
| getMUBUFIns<addrKindCopy, [vdataClassCopy]>.ret, |
| " $vdata, " # getMUBUFAsmOps<addrKindCopy>.ret # "$glc$slc$tfe", |
| pattern>, |
| MUBUF_SetupAddr<addrKindCopy> { |
| let PseudoInstr = opName # "_" # getAddrName<addrKindCopy>.ret; |
| let mayLoad = 0; |
| let mayStore = 1; |
| let maybeAtomic = 1; |
| } |
| |
| multiclass MUBUF_Pseudo_Stores<string opName, RegisterClass vdataClass, |
| ValueType store_vt = i32, |
| SDPatternOperator st = null_frag> { |
| |
| def _OFFSET : MUBUF_Store_Pseudo <opName, BUFAddrKind.Offset, vdataClass, |
| [(st store_vt:$vdata, (MUBUFOffset v4i32:$srsrc, i32:$soffset, |
| i16:$offset, i1:$glc, i1:$slc, i1:$tfe))]>, |
| MUBUFAddr64Table<0, NAME>; |
| |
| def _ADDR64 : MUBUF_Store_Pseudo <opName, BUFAddrKind.Addr64, vdataClass, |
| [(st store_vt:$vdata, (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, |
| i16:$offset, i1:$glc, i1:$slc, i1:$tfe))]>, |
| MUBUFAddr64Table<1, NAME>; |
| |
| def _OFFEN : MUBUF_Store_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN : MUBUF_Store_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN : MUBUF_Store_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| |
| let DisableWQM = 1 in { |
| def _OFFSET_exact : MUBUF_Store_Pseudo <opName, BUFAddrKind.Offset, vdataClass>; |
| def _OFFEN_exact : MUBUF_Store_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN_exact : MUBUF_Store_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN_exact : MUBUF_Store_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| } |
| } |
| |
| class MUBUF_Pseudo_Store_Lds<string opName> |
| : MUBUF_Pseudo<opName, |
| (outs), |
| (ins SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc), |
| " $srsrc, $soffset$offset lds$glc$slc"> { |
| let mayLoad = 0; |
| let mayStore = 1; |
| let maybeAtomic = 1; |
| |
| let has_vdata = 0; |
| let has_vaddr = 0; |
| let has_tfe = 0; |
| let lds = 1; |
| |
| let Uses = [EXEC, M0]; |
| let AsmMatchConverter = "cvtMubufLds"; |
| } |
| |
| class getMUBUFAtomicInsDA<RegisterClass vdataClass, bit vdata_in, |
| list<RegisterClass> vaddrList=[]> { |
| RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); |
| dag ret = !if(vdata_in, |
| !if(!empty(vaddrList), |
| (ins vdataClass:$vdata_in, |
| SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc), |
| (ins vdataClass:$vdata_in, vaddrClass:$vaddr, |
| SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc) |
| ), |
| !if(!empty(vaddrList), |
| (ins vdataClass:$vdata, |
| SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc), |
| (ins vdataClass:$vdata, vaddrClass:$vaddr, |
| SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc) |
| )); |
| } |
| |
| class getMUBUFAtomicIns<int addrKind, |
| RegisterClass vdataClass, |
| bit vdata_in, |
| // Workaround bug bz30254 |
| RegisterClass vdataClassCopy=vdataClass> { |
| dag ret = |
| !if(!eq(addrKind, BUFAddrKind.Offset), |
| getMUBUFAtomicInsDA<vdataClassCopy, vdata_in>.ret, |
| !if(!eq(addrKind, BUFAddrKind.OffEn), |
| getMUBUFAtomicInsDA<vdataClassCopy, vdata_in, [VGPR_32]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.IdxEn), |
| getMUBUFAtomicInsDA<vdataClassCopy, vdata_in, [VGPR_32]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.BothEn), |
| getMUBUFAtomicInsDA<vdataClassCopy, vdata_in, [VReg_64]>.ret, |
| !if(!eq(addrKind, BUFAddrKind.Addr64), |
| getMUBUFAtomicInsDA<vdataClassCopy, vdata_in, [VReg_64]>.ret, |
| (ins)))))); |
| } |
| |
| class MUBUF_Atomic_Pseudo<string opName, |
| int addrKind, |
| dag outs, |
| dag ins, |
| string asmOps, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind> |
| : MUBUF_Pseudo<opName, outs, ins, asmOps, pattern>, |
| MUBUF_SetupAddr<addrKindCopy> { |
| let mayStore = 1; |
| let mayLoad = 1; |
| let hasPostISelHook = 1; |
| let hasSideEffects = 1; |
| let DisableWQM = 1; |
| let has_glc = 0; |
| let has_tfe = 0; |
| let maybeAtomic = 1; |
| } |
| |
| class MUBUF_AtomicNoRet_Pseudo<string opName, int addrKind, |
| RegisterClass vdataClass, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind, |
| RegisterClass vdataClassCopy = vdataClass> |
| : MUBUF_Atomic_Pseudo<opName, addrKindCopy, |
| (outs), |
| getMUBUFAtomicIns<addrKindCopy, vdataClassCopy, 0>.ret, |
| " $vdata, " # getMUBUFAsmOps<addrKindCopy>.ret # "$slc", |
| pattern>, |
| AtomicNoRet<opName # "_" # getAddrName<addrKindCopy>.ret, 0> { |
| let PseudoInstr = opName # "_" # getAddrName<addrKindCopy>.ret; |
| let glc_value = 0; |
| let AsmMatchConverter = "cvtMubufAtomic"; |
| } |
| |
| class MUBUF_AtomicRet_Pseudo<string opName, int addrKind, |
| RegisterClass vdataClass, |
| list<dag> pattern=[], |
| // Workaround bug bz30254 |
| int addrKindCopy = addrKind, |
| RegisterClass vdataClassCopy = vdataClass> |
| : MUBUF_Atomic_Pseudo<opName, addrKindCopy, |
| (outs vdataClassCopy:$vdata), |
| getMUBUFAtomicIns<addrKindCopy, vdataClassCopy, 1>.ret, |
| " $vdata, " # getMUBUFAsmOps<addrKindCopy>.ret # " glc$slc", |
| pattern>, |
| AtomicNoRet<opName # "_" # getAddrName<addrKindCopy>.ret, 1> { |
| let PseudoInstr = opName # "_rtn_" # getAddrName<addrKindCopy>.ret; |
| let glc_value = 1; |
| let Constraints = "$vdata = $vdata_in"; |
| let DisableEncoding = "$vdata_in"; |
| let AsmMatchConverter = "cvtMubufAtomicReturn"; |
| } |
| |
| multiclass MUBUF_Pseudo_Atomics <string opName, |
| RegisterClass vdataClass, |
| ValueType vdataType, |
| SDPatternOperator atomic> { |
| |
| def _OFFSET : MUBUF_AtomicNoRet_Pseudo <opName, BUFAddrKind.Offset, vdataClass>, |
| MUBUFAddr64Table <0, NAME>; |
| def _ADDR64 : MUBUF_AtomicNoRet_Pseudo <opName, BUFAddrKind.Addr64, vdataClass>, |
| MUBUFAddr64Table <1, NAME>; |
| def _OFFEN : MUBUF_AtomicNoRet_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN : MUBUF_AtomicNoRet_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN : MUBUF_AtomicNoRet_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| |
| def _OFFSET_RTN : MUBUF_AtomicRet_Pseudo <opName, BUFAddrKind.Offset, vdataClass, |
| [(set vdataType:$vdata, |
| (atomic (MUBUFOffsetAtomic v4i32:$srsrc, i32:$soffset, i16:$offset, i1:$slc), |
| vdataType:$vdata_in))]>, |
| MUBUFAddr64Table <0, NAME # "_RTN">; |
| |
| def _ADDR64_RTN : MUBUF_AtomicRet_Pseudo <opName, BUFAddrKind.Addr64, vdataClass, |
| [(set vdataType:$vdata, |
| (atomic (MUBUFAddr64Atomic v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, i1:$slc), |
| vdataType:$vdata_in))]>, |
| MUBUFAddr64Table <1, NAME # "_RTN">; |
| |
| def _OFFEN_RTN : MUBUF_AtomicRet_Pseudo <opName, BUFAddrKind.OffEn, vdataClass>; |
| def _IDXEN_RTN : MUBUF_AtomicRet_Pseudo <opName, BUFAddrKind.IdxEn, vdataClass>; |
| def _BOTHEN_RTN : MUBUF_AtomicRet_Pseudo <opName, BUFAddrKind.BothEn, vdataClass>; |
| } |
| |
| |
| //===----------------------------------------------------------------------===// |
| // MUBUF Instructions |
| //===----------------------------------------------------------------------===// |
| |
| defm BUFFER_LOAD_FORMAT_X : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_format_x", VGPR_32 |
| >; |
| defm BUFFER_LOAD_FORMAT_XY : MUBUF_Pseudo_Loads < |
| "buffer_load_format_xy", VReg_64 |
| >; |
| defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Pseudo_Loads < |
| "buffer_load_format_xyz", VReg_96 |
| >; |
| defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Pseudo_Loads < |
| "buffer_load_format_xyzw", VReg_128 |
| >; |
| defm BUFFER_STORE_FORMAT_X : MUBUF_Pseudo_Stores < |
| "buffer_store_format_x", VGPR_32 |
| >; |
| defm BUFFER_STORE_FORMAT_XY : MUBUF_Pseudo_Stores < |
| "buffer_store_format_xy", VReg_64 |
| >; |
| defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Pseudo_Stores < |
| "buffer_store_format_xyz", VReg_96 |
| >; |
| defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Pseudo_Stores < |
| "buffer_store_format_xyzw", VReg_128 |
| >; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem, D16Buf = 1 in { |
| defm BUFFER_LOAD_FORMAT_D16_X_gfx80 : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_x", VGPR_32 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XY_gfx80 : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xy", VReg_64 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xyz", VReg_96 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xyzw", VReg_128 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_X_gfx80 : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_x", VGPR_32 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XY_gfx80 : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xy", VReg_64 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xyz", VReg_96 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xyzw", VReg_128 |
| >; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem, D16Buf = 1 in { |
| defm BUFFER_LOAD_FORMAT_D16_X : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_x", VGPR_32 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XY : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xy", VGPR_32 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XYZ : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xyz", VReg_64 |
| >; |
| defm BUFFER_LOAD_FORMAT_D16_XYZW : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_xyzw", VReg_64 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_X : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_x", VGPR_32 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XY : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xy", VGPR_32 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XYZ : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xyz", VReg_64 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_XYZW : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_xyzw", VReg_64 |
| >; |
| } // End HasPackedD16VMem. |
| |
| defm BUFFER_LOAD_UBYTE : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_ubyte", VGPR_32, i32, mubuf_az_extloadi8 |
| >; |
| defm BUFFER_LOAD_SBYTE : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_sbyte", VGPR_32, i32, mubuf_sextloadi8 |
| >; |
| defm BUFFER_LOAD_USHORT : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_ushort", VGPR_32, i32, mubuf_az_extloadi16 |
| >; |
| defm BUFFER_LOAD_SSHORT : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_sshort", VGPR_32, i32, mubuf_sextloadi16 |
| >; |
| defm BUFFER_LOAD_DWORD : MUBUF_Pseudo_Loads_Lds < |
| "buffer_load_dword", VGPR_32, i32, mubuf_load |
| >; |
| defm BUFFER_LOAD_DWORDX2 : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx2", VReg_64, v2i32, mubuf_load |
| >; |
| defm BUFFER_LOAD_DWORDX3 : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx3", VReg_96, untyped, mubuf_load |
| >; |
| defm BUFFER_LOAD_DWORDX4 : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx4", VReg_128, v4i32, mubuf_load |
| >; |
| |
| // This is not described in AMD documentation, |
| // but 'lds' versions of these opcodes are available |
| // in at least GFX8+ chips. See Bug 37653. |
| let SubtargetPredicate = isVI in { |
| defm BUFFER_LOAD_DWORDX2_LDS : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx2", VReg_64, v2i32, null_frag, 0, 1 |
| >; |
| defm BUFFER_LOAD_DWORDX3_LDS : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx3", VReg_96, untyped, null_frag, 0, 1 |
| >; |
| defm BUFFER_LOAD_DWORDX4_LDS : MUBUF_Pseudo_Loads < |
| "buffer_load_dwordx4", VReg_128, v4i32, null_frag, 0, 1 |
| >; |
| } |
| |
| defm BUFFER_STORE_BYTE : MUBUF_Pseudo_Stores < |
| "buffer_store_byte", VGPR_32, i32, truncstorei8_global |
| >; |
| defm BUFFER_STORE_SHORT : MUBUF_Pseudo_Stores < |
| "buffer_store_short", VGPR_32, i32, truncstorei16_global |
| >; |
| defm BUFFER_STORE_DWORD : MUBUF_Pseudo_Stores < |
| "buffer_store_dword", VGPR_32, i32, store_global |
| >; |
| defm BUFFER_STORE_DWORDX2 : MUBUF_Pseudo_Stores < |
| "buffer_store_dwordx2", VReg_64, v2i32, store_global |
| >; |
| defm BUFFER_STORE_DWORDX3 : MUBUF_Pseudo_Stores < |
| "buffer_store_dwordx3", VReg_96, untyped, store_global |
| >; |
| defm BUFFER_STORE_DWORDX4 : MUBUF_Pseudo_Stores < |
| "buffer_store_dwordx4", VReg_128, v4i32, store_global |
| >; |
| defm BUFFER_ATOMIC_SWAP : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_swap", VGPR_32, i32, atomic_swap_global |
| >; |
| defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_cmpswap", VReg_64, v2i32, null_frag |
| >; |
| defm BUFFER_ATOMIC_ADD : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_add", VGPR_32, i32, atomic_add_global |
| >; |
| defm BUFFER_ATOMIC_SUB : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_sub", VGPR_32, i32, atomic_sub_global |
| >; |
| defm BUFFER_ATOMIC_SMIN : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_smin", VGPR_32, i32, atomic_min_global |
| >; |
| defm BUFFER_ATOMIC_UMIN : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_umin", VGPR_32, i32, atomic_umin_global |
| >; |
| defm BUFFER_ATOMIC_SMAX : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_smax", VGPR_32, i32, atomic_max_global |
| >; |
| defm BUFFER_ATOMIC_UMAX : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_umax", VGPR_32, i32, atomic_umax_global |
| >; |
| defm BUFFER_ATOMIC_AND : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_and", VGPR_32, i32, atomic_and_global |
| >; |
| defm BUFFER_ATOMIC_OR : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_or", VGPR_32, i32, atomic_or_global |
| >; |
| defm BUFFER_ATOMIC_XOR : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_xor", VGPR_32, i32, atomic_xor_global |
| >; |
| defm BUFFER_ATOMIC_INC : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_inc", VGPR_32, i32, atomic_inc_global |
| >; |
| defm BUFFER_ATOMIC_DEC : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_dec", VGPR_32, i32, atomic_dec_global |
| >; |
| defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_swap_x2", VReg_64, i64, atomic_swap_global |
| >; |
| defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_cmpswap_x2", VReg_128, v2i64, null_frag |
| >; |
| defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_add_x2", VReg_64, i64, atomic_add_global |
| >; |
| defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_sub_x2", VReg_64, i64, atomic_sub_global |
| >; |
| defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_smin_x2", VReg_64, i64, atomic_min_global |
| >; |
| defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_umin_x2", VReg_64, i64, atomic_umin_global |
| >; |
| defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_smax_x2", VReg_64, i64, atomic_max_global |
| >; |
| defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_umax_x2", VReg_64, i64, atomic_umax_global |
| >; |
| defm BUFFER_ATOMIC_AND_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_and_x2", VReg_64, i64, atomic_and_global |
| >; |
| defm BUFFER_ATOMIC_OR_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_or_x2", VReg_64, i64, atomic_or_global |
| >; |
| defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_xor_x2", VReg_64, i64, atomic_xor_global |
| >; |
| defm BUFFER_ATOMIC_INC_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_inc_x2", VReg_64, i64, atomic_inc_global |
| >; |
| defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Pseudo_Atomics < |
| "buffer_atomic_dec_x2", VReg_64, i64, atomic_dec_global |
| >; |
| |
| let SubtargetPredicate = isVI in { |
| def BUFFER_STORE_LDS_DWORD : MUBUF_Pseudo_Store_Lds <"buffer_store_lds_dword">; |
| } |
| |
| let SubtargetPredicate = isSI in { // isn't on CI & VI |
| /* |
| defm BUFFER_ATOMIC_RSUB : MUBUF_Pseudo_Atomics <"buffer_atomic_rsub">; |
| defm BUFFER_ATOMIC_FCMPSWAP : MUBUF_Pseudo_Atomics <"buffer_atomic_fcmpswap">; |
| defm BUFFER_ATOMIC_FMIN : MUBUF_Pseudo_Atomics <"buffer_atomic_fmin">; |
| defm BUFFER_ATOMIC_FMAX : MUBUF_Pseudo_Atomics <"buffer_atomic_fmax">; |
| defm BUFFER_ATOMIC_RSUB_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_rsub_x2">; |
| defm BUFFER_ATOMIC_FCMPSWAP_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fcmpswap_x2">; |
| defm BUFFER_ATOMIC_FMIN_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fmin_x2">; |
| defm BUFFER_ATOMIC_FMAX_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fmax_x2">; |
| */ |
| |
| def BUFFER_WBINVL1_SC : MUBUF_Invalidate <"buffer_wbinvl1_sc", |
| int_amdgcn_buffer_wbinvl1_sc>; |
| } |
| |
| let SubtargetPredicate = HasD16LoadStore in { |
| |
| defm BUFFER_LOAD_UBYTE_D16 : MUBUF_Pseudo_Loads < |
| "buffer_load_ubyte_d16", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_LOAD_UBYTE_D16_HI : MUBUF_Pseudo_Loads < |
| "buffer_load_ubyte_d16_hi", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_LOAD_SBYTE_D16 : MUBUF_Pseudo_Loads < |
| "buffer_load_sbyte_d16", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_LOAD_SBYTE_D16_HI : MUBUF_Pseudo_Loads < |
| "buffer_load_sbyte_d16_hi", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_LOAD_SHORT_D16 : MUBUF_Pseudo_Loads < |
| "buffer_load_short_d16", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_LOAD_SHORT_D16_HI : MUBUF_Pseudo_Loads < |
| "buffer_load_short_d16_hi", VGPR_32, i32, null_frag, 1 |
| >; |
| |
| defm BUFFER_STORE_BYTE_D16_HI : MUBUF_Pseudo_Stores < |
| "buffer_store_byte_d16_hi", VGPR_32, i32 |
| >; |
| |
| defm BUFFER_STORE_SHORT_D16_HI : MUBUF_Pseudo_Stores < |
| "buffer_store_short_d16_hi", VGPR_32, i32 |
| >; |
| |
| defm BUFFER_LOAD_FORMAT_D16_HI_X : MUBUF_Pseudo_Loads < |
| "buffer_load_format_d16_hi_x", VGPR_32 |
| >; |
| defm BUFFER_STORE_FORMAT_D16_HI_X : MUBUF_Pseudo_Stores < |
| "buffer_store_format_d16_hi_x", VGPR_32 |
| >; |
| |
| } // End HasD16LoadStore |
| |
| def BUFFER_WBINVL1 : MUBUF_Invalidate <"buffer_wbinvl1", |
| int_amdgcn_buffer_wbinvl1>; |
| |
| //===----------------------------------------------------------------------===// |
| // MTBUF Instructions |
| //===----------------------------------------------------------------------===// |
| |
| defm TBUFFER_LOAD_FORMAT_X : MTBUF_Pseudo_Loads <"tbuffer_load_format_x", VGPR_32>; |
| defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Pseudo_Loads <"tbuffer_load_format_xy", VReg_64>; |
| defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Pseudo_Loads <"tbuffer_load_format_xyz", VReg_128>; |
| defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Pseudo_Loads <"tbuffer_load_format_xyzw", VReg_128>; |
| defm TBUFFER_STORE_FORMAT_X : MTBUF_Pseudo_Stores <"tbuffer_store_format_x", VGPR_32>; |
| defm TBUFFER_STORE_FORMAT_XY : MTBUF_Pseudo_Stores <"tbuffer_store_format_xy", VReg_64>; |
| defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Pseudo_Stores <"tbuffer_store_format_xyz", VReg_128>; |
| defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Pseudo_Stores <"tbuffer_store_format_xyzw", VReg_128>; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem, D16Buf = 1 in { |
| defm TBUFFER_LOAD_FORMAT_D16_X_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_x", VGPR_32>; |
| defm TBUFFER_LOAD_FORMAT_D16_XY_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xy", VReg_64>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyz", VReg_96>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyzw", VReg_128>; |
| defm TBUFFER_STORE_FORMAT_D16_X_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_x", VGPR_32>; |
| defm TBUFFER_STORE_FORMAT_D16_XY_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xy", VReg_64>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyz", VReg_96>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyzw", VReg_128>; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem, D16Buf = 1 in { |
| defm TBUFFER_LOAD_FORMAT_D16_X : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_x", VGPR_32>; |
| defm TBUFFER_LOAD_FORMAT_D16_XY : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xy", VGPR_32>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZ : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyz", VReg_64>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZW : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyzw", VReg_64>; |
| defm TBUFFER_STORE_FORMAT_D16_X : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_x", VGPR_32>; |
| defm TBUFFER_STORE_FORMAT_D16_XY : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xy", VGPR_32>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZ : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyz", VReg_64>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZW : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyzw", VReg_64>; |
| } // End HasPackedD16VMem. |
| |
| let SubtargetPredicate = isCIVI in { |
| |
| //===----------------------------------------------------------------------===// |
| // Instruction definitions for CI and newer. |
| //===----------------------------------------------------------------------===// |
| // Remaining instructions: |
| // BUFFER_LOAD_DWORDX3 |
| // BUFFER_STORE_DWORDX3 |
| |
| def BUFFER_WBINVL1_VOL : MUBUF_Invalidate <"buffer_wbinvl1_vol", |
| int_amdgcn_buffer_wbinvl1_vol>; |
| |
| } // End let SubtargetPredicate = isCIVI |
| |
| //===----------------------------------------------------------------------===// |
| // MUBUF Patterns |
| //===----------------------------------------------------------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| // buffer_load/store_format patterns |
| //===----------------------------------------------------------------------===// |
| |
| multiclass MUBUF_LoadIntrinsicPat<SDPatternOperator name, ValueType vt, |
| string opcode> { |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, 0, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$glc, imm:$slc)), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFSET) $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$glc, imm:$slc)), |
| (!cast<MUBUF_Pseudo>(opcode # _IDXEN) $vindex, $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, 0, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$glc, imm:$slc)), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFEN) $voffset, $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$glc, imm:$slc)), |
| (!cast<MUBUF_Pseudo>(opcode # _BOTHEN) |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| } |
| |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format, f32, "BUFFER_LOAD_FORMAT_X">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format, v2f32, "BUFFER_LOAD_FORMAT_XY">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format, v4f32, "BUFFER_LOAD_FORMAT_XYZW">; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, f16, "BUFFER_LOAD_FORMAT_D16_X_gfx80">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, v2i32, "BUFFER_LOAD_FORMAT_D16_XY_gfx80">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, v4i32, "BUFFER_LOAD_FORMAT_D16_XYZW_gfx80">; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, f16, "BUFFER_LOAD_FORMAT_D16_X">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, v2f16, "BUFFER_LOAD_FORMAT_D16_XY">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load_format_d16, v4f16, "BUFFER_LOAD_FORMAT_D16_XYZW">; |
| } // End HasPackedD16VMem. |
| |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load, f32, "BUFFER_LOAD_DWORD">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load, v2f32, "BUFFER_LOAD_DWORDX2">; |
| defm : MUBUF_LoadIntrinsicPat<SIbuffer_load, v4f32, "BUFFER_LOAD_DWORDX4">; |
| |
| multiclass MUBUF_StoreIntrinsicPat<SDPatternOperator name, ValueType vt, |
| string opcode> { |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$glc, imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFSET_exact) $vdata, $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$glc, imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _IDXEN_exact) $vdata, $vindex, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i1imm $glc), |
| (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$glc, imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFEN_exact) $vdata, $voffset, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i1imm $glc), |
| (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$glc, imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _BOTHEN_exact) |
| $vdata, |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), |
| (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| } |
| |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format, f32, "BUFFER_STORE_FORMAT_X">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format, v2f32, "BUFFER_STORE_FORMAT_XY">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format, v4f32, "BUFFER_STORE_FORMAT_XYZW">; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, f16, "BUFFER_STORE_FORMAT_D16_X_gfx80">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, v2i32, "BUFFER_STORE_FORMAT_D16_XY_gfx80">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, v4i32, "BUFFER_STORE_FORMAT_D16_XYZW_gfx80">; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, f16, "BUFFER_STORE_FORMAT_D16_X">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, v2f16, "BUFFER_STORE_FORMAT_D16_XY">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store_format_d16, v4f16, "BUFFER_STORE_FORMAT_D16_XYZW">; |
| } // End HasPackedD16VMem. |
| |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store, f32, "BUFFER_STORE_DWORD">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store, v2f32, "BUFFER_STORE_DWORDX2">; |
| defm : MUBUF_StoreIntrinsicPat<SIbuffer_store, v4f32, "BUFFER_STORE_DWORDX4">; |
| |
| //===----------------------------------------------------------------------===// |
| // buffer_atomic patterns |
| //===----------------------------------------------------------------------===// |
| |
| multiclass BufferAtomicPatterns<SDPatternOperator name, string opcode> { |
| def : GCNPat< |
| (name i32:$vdata_in, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFSET_RTN) $vdata_in, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i1imm $slc)) |
| >; |
| |
| def : GCNPat< |
| (name i32:$vdata_in, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _IDXEN_RTN) $vdata_in, $vindex, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i1imm $slc)) |
| >; |
| |
| def : GCNPat< |
| (name i32:$vdata_in, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _OFFEN_RTN) $vdata_in, $voffset, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i1imm $slc)) |
| >; |
| |
| def : GCNPat< |
| (name i32:$vdata_in, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$slc), |
| (!cast<MUBUF_Pseudo>(opcode # _BOTHEN_RTN) |
| $vdata_in, |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)) |
| >; |
| } |
| |
| defm : BufferAtomicPatterns<SIbuffer_atomic_swap, "BUFFER_ATOMIC_SWAP">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_add, "BUFFER_ATOMIC_ADD">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_sub, "BUFFER_ATOMIC_SUB">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_smin, "BUFFER_ATOMIC_SMIN">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_umin, "BUFFER_ATOMIC_UMIN">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_smax, "BUFFER_ATOMIC_SMAX">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_umax, "BUFFER_ATOMIC_UMAX">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_and, "BUFFER_ATOMIC_AND">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_or, "BUFFER_ATOMIC_OR">; |
| defm : BufferAtomicPatterns<SIbuffer_atomic_xor, "BUFFER_ATOMIC_XOR">; |
| |
| def : GCNPat< |
| (SIbuffer_atomic_cmpswap |
| i32:$data, i32:$cmp, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$slc), |
| (EXTRACT_SUBREG |
| (BUFFER_ATOMIC_CMPSWAP_OFFSET_RTN |
| (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), |
| sub0) |
| >; |
| |
| def : GCNPat< |
| (SIbuffer_atomic_cmpswap |
| i32:$data, i32:$cmp, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), |
| imm:$slc), |
| (EXTRACT_SUBREG |
| (BUFFER_ATOMIC_CMPSWAP_IDXEN_RTN |
| (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), |
| $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), |
| sub0) |
| >; |
| |
| def : GCNPat< |
| (SIbuffer_atomic_cmpswap |
| i32:$data, i32:$cmp, v4i32:$rsrc, 0, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$slc), |
| (EXTRACT_SUBREG |
| (BUFFER_ATOMIC_CMPSWAP_OFFEN_RTN |
| (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), |
| $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), |
| sub0) |
| >; |
| |
| def : GCNPat< |
| (SIbuffer_atomic_cmpswap |
| i32:$data, i32:$cmp, v4i32:$rsrc, i32:$vindex, |
| (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), |
| imm:$slc), |
| (EXTRACT_SUBREG |
| (BUFFER_ATOMIC_CMPSWAP_BOTHEN_RTN |
| (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), |
| sub0) |
| >; |
| |
| |
| class MUBUFLoad_PatternADDR64 <MUBUF_Pseudo Instr_ADDR64, ValueType vt, |
| PatFrag constant_ld> : GCNPat < |
| (vt (constant_ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, |
| i16:$offset, i1:$glc, i1:$slc, i1:$tfe))), |
| (Instr_ADDR64 $vaddr, $srsrc, $soffset, $offset, $glc, $slc, $tfe) |
| >; |
| |
| multiclass MUBUFLoad_Atomic_Pattern <MUBUF_Pseudo Instr_ADDR64, MUBUF_Pseudo Instr_OFFSET, |
| ValueType vt, PatFrag atomic_ld> { |
| def : GCNPat < |
| (vt (atomic_ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, |
| i16:$offset, i1:$slc))), |
| (Instr_ADDR64 $vaddr, $srsrc, $soffset, $offset, 0, $slc, 0) |
| >; |
| |
| def : GCNPat < |
| (vt (atomic_ld (MUBUFOffsetNoGLC v4i32:$rsrc, i32:$soffset, i16:$offset))), |
| (Instr_OFFSET $rsrc, $soffset, (as_i16imm $offset), 0, 0, 0) |
| >; |
| } |
| |
| let SubtargetPredicate = isSICI in { |
| def : MUBUFLoad_PatternADDR64 <BUFFER_LOAD_SBYTE_ADDR64, i32, sextloadi8_constant>; |
| def : MUBUFLoad_PatternADDR64 <BUFFER_LOAD_UBYTE_ADDR64, i32, az_extloadi8_constant>; |
| def : MUBUFLoad_PatternADDR64 <BUFFER_LOAD_SSHORT_ADDR64, i32, sextloadi16_constant>; |
| def : MUBUFLoad_PatternADDR64 <BUFFER_LOAD_USHORT_ADDR64, i32, az_extloadi16_constant>; |
| |
| defm : MUBUFLoad_Atomic_Pattern <BUFFER_LOAD_DWORD_ADDR64, BUFFER_LOAD_DWORD_OFFSET, i32, mubuf_load_atomic>; |
| defm : MUBUFLoad_Atomic_Pattern <BUFFER_LOAD_DWORDX2_ADDR64, BUFFER_LOAD_DWORDX2_OFFSET, i64, mubuf_load_atomic>; |
| } // End SubtargetPredicate = isSICI |
| |
| multiclass MUBUFLoad_Pattern <MUBUF_Pseudo Instr_OFFSET, ValueType vt, |
| PatFrag ld> { |
| |
| def : GCNPat < |
| (vt (ld (MUBUFOffset v4i32:$srsrc, i32:$soffset, |
| i16:$offset, i1:$glc, i1:$slc, i1:$tfe))), |
| (Instr_OFFSET $srsrc, $soffset, $offset, $glc, $slc, $tfe) |
| >; |
| } |
| |
| let OtherPredicates = [Has16BitInsts] in { |
| |
| defm : MUBUFLoad_Pattern <BUFFER_LOAD_SBYTE_OFFSET, i16, sextloadi8_constant>; |
| defm : MUBUFLoad_Pattern <BUFFER_LOAD_UBYTE_OFFSET, i16, az_extloadi8_constant>; |
| defm : MUBUFLoad_Pattern <BUFFER_LOAD_SBYTE_OFFSET, i16, mubuf_sextloadi8>; |
| defm : MUBUFLoad_Pattern <BUFFER_LOAD_UBYTE_OFFSET, i16, mubuf_az_extloadi8>; |
| |
| defm : MUBUFLoad_Pattern <BUFFER_LOAD_USHORT_OFFSET, i16, mubuf_load>; |
| |
| } // End OtherPredicates = [Has16BitInsts] |
| |
| multiclass MUBUFScratchLoadPat <MUBUF_Pseudo InstrOffen, |
| MUBUF_Pseudo InstrOffset, |
| ValueType vt, PatFrag ld> { |
| def : GCNPat < |
| (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset))), |
| (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0) |
| >; |
| |
| def : GCNPat < |
| (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))), |
| (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0) |
| >; |
| } |
| |
| // XXX - Is it possible to have a complex pattern in a PatFrag? |
| multiclass MUBUFScratchLoadPat_Hi16 <MUBUF_Pseudo InstrOffen, |
| MUBUF_Pseudo InstrOffset, |
| ValueType vt, PatFrag ld> { |
| def : GCNPat < |
| (build_vector vt:$lo, (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset)))), |
| (v2i16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $lo)) |
| >; |
| |
| def : GCNPat < |
| (build_vector f16:$lo, (f16 (bitconvert (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset)))))), |
| (v2f16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $lo)) |
| >; |
| |
| |
| def : GCNPat < |
| (build_vector vt:$lo, (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset)))), |
| (v2i16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $lo)) |
| >; |
| |
| def : GCNPat < |
| (build_vector f16:$lo, (f16 (bitconvert (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset)))))), |
| (v2f16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $lo)) |
| >; |
| } |
| |
| multiclass MUBUFScratchLoadPat_Lo16 <MUBUF_Pseudo InstrOffen, |
| MUBUF_Pseudo InstrOffset, |
| ValueType vt, PatFrag ld> { |
| def : GCNPat < |
| (build_vector (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset))), |
| (vt (Hi16Elt vt:$hi))), |
| (v2i16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $hi)) |
| >; |
| |
| def : GCNPat < |
| (build_vector (f16 (bitconvert (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset))))), |
| (f16 (Hi16Elt f16:$hi))), |
| (v2f16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $hi)) |
| >; |
| |
| def : GCNPat < |
| (build_vector (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))), |
| (vt (Hi16Elt vt:$hi))), |
| (v2i16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $hi)) |
| >; |
| |
| def : GCNPat < |
| (build_vector (f16 (bitconvert (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))))), |
| (f16 (Hi16Elt f16:$hi))), |
| (v2f16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $hi)) |
| >; |
| } |
| |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_SBYTE_OFFEN, BUFFER_LOAD_SBYTE_OFFSET, i32, sextloadi8_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_UBYTE_OFFEN, BUFFER_LOAD_UBYTE_OFFSET, i32, az_extloadi8_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_SBYTE_OFFEN, BUFFER_LOAD_SBYTE_OFFSET, i16, sextloadi8_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_UBYTE_OFFEN, BUFFER_LOAD_UBYTE_OFFSET, i16, az_extloadi8_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_SSHORT_OFFEN, BUFFER_LOAD_SSHORT_OFFSET, i32, sextloadi16_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_USHORT_OFFEN, BUFFER_LOAD_USHORT_OFFSET, i32, az_extloadi16_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_USHORT_OFFEN, BUFFER_LOAD_USHORT_OFFSET, i16, load_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_DWORD_OFFEN, BUFFER_LOAD_DWORD_OFFSET, i32, load_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_DWORDX2_OFFEN, BUFFER_LOAD_DWORDX2_OFFSET, v2i32, load_private>; |
| defm : MUBUFScratchLoadPat <BUFFER_LOAD_DWORDX4_OFFEN, BUFFER_LOAD_DWORDX4_OFFSET, v4i32, load_private>; |
| |
| let OtherPredicates = [D16PreservesUnusedBits] in { |
| defm : MUBUFScratchLoadPat_Hi16<BUFFER_LOAD_SHORT_D16_HI_OFFEN, BUFFER_LOAD_SHORT_D16_HI_OFFSET, i16, load_private>; |
| defm : MUBUFScratchLoadPat_Hi16<BUFFER_LOAD_UBYTE_D16_HI_OFFEN, BUFFER_LOAD_UBYTE_D16_HI_OFFSET, i16, az_extloadi8_private>; |
| defm : MUBUFScratchLoadPat_Hi16<BUFFER_LOAD_SBYTE_D16_HI_OFFEN, BUFFER_LOAD_SBYTE_D16_HI_OFFSET, i16, sextloadi8_private>; |
| |
| defm : MUBUFScratchLoadPat_Lo16<BUFFER_LOAD_SHORT_D16_OFFEN, BUFFER_LOAD_SHORT_D16_OFFSET, i16, load_private>; |
| defm : MUBUFScratchLoadPat_Lo16<BUFFER_LOAD_UBYTE_D16_OFFEN, BUFFER_LOAD_UBYTE_D16_OFFSET, i16, az_extloadi8_private>; |
| defm : MUBUFScratchLoadPat_Lo16<BUFFER_LOAD_SBYTE_D16_OFFEN, BUFFER_LOAD_SBYTE_D16_OFFSET, i16, sextloadi8_private>; |
| } |
| |
| // BUFFER_LOAD_DWORD*, addr64=0 |
| multiclass MUBUF_Load_Dword <ValueType vt, |
| MUBUF_Pseudo offset, |
| MUBUF_Pseudo offen, |
| MUBUF_Pseudo idxen, |
| MUBUF_Pseudo bothen> { |
| |
| def : GCNPat < |
| (vt (int_SI_buffer_load_dword v4i32:$rsrc, (i32 imm), i32:$soffset, |
| imm:$offset, 0, 0, imm:$glc, imm:$slc, |
| imm:$tfe)), |
| (offset $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), |
| (as_i1imm $slc), (as_i1imm $tfe)) |
| >; |
| |
| def : GCNPat < |
| (vt (int_SI_buffer_load_dword v4i32:$rsrc, i32:$vaddr, i32:$soffset, |
| imm:$offset, 1, 0, imm:$glc, imm:$slc, |
| imm:$tfe)), |
| (offen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), |
| (as_i1imm $tfe)) |
| >; |
| |
| def : GCNPat < |
| (vt (int_SI_buffer_load_dword v4i32:$rsrc, i32:$vaddr, i32:$soffset, |
| imm:$offset, 0, 1, imm:$glc, imm:$slc, |
| imm:$tfe)), |
| (idxen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), |
| (as_i1imm $slc), (as_i1imm $tfe)) |
| >; |
| |
| def : GCNPat < |
| (vt (int_SI_buffer_load_dword v4i32:$rsrc, v2i32:$vaddr, i32:$soffset, |
| imm:$offset, 1, 1, imm:$glc, imm:$slc, |
| imm:$tfe)), |
| (bothen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), |
| (as_i1imm $tfe)) |
| >; |
| } |
| |
| defm : MUBUF_Load_Dword <i32, BUFFER_LOAD_DWORD_OFFSET, BUFFER_LOAD_DWORD_OFFEN, |
| BUFFER_LOAD_DWORD_IDXEN, BUFFER_LOAD_DWORD_BOTHEN>; |
| defm : MUBUF_Load_Dword <v2i32, BUFFER_LOAD_DWORDX2_OFFSET, BUFFER_LOAD_DWORDX2_OFFEN, |
| BUFFER_LOAD_DWORDX2_IDXEN, BUFFER_LOAD_DWORDX2_BOTHEN>; |
| defm : MUBUF_Load_Dword <v4i32, BUFFER_LOAD_DWORDX4_OFFSET, BUFFER_LOAD_DWORDX4_OFFEN, |
| BUFFER_LOAD_DWORDX4_IDXEN, BUFFER_LOAD_DWORDX4_BOTHEN>; |
| |
| multiclass MUBUFStore_Atomic_Pattern <MUBUF_Pseudo Instr_ADDR64, MUBUF_Pseudo Instr_OFFSET, |
| ValueType vt, PatFrag atomic_st> { |
| // Store follows atomic op convention so address is forst |
| def : GCNPat < |
| (atomic_st (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, |
| i16:$offset, i1:$slc), vt:$val), |
| (Instr_ADDR64 $val, $vaddr, $srsrc, $soffset, $offset, 0, $slc, 0) |
| >; |
| |
| def : GCNPat < |
| (atomic_st (MUBUFOffsetNoGLC v4i32:$rsrc, i32:$soffset, i16:$offset), vt:$val), |
| (Instr_OFFSET $val, $rsrc, $soffset, (as_i16imm $offset), 0, 0, 0) |
| >; |
| } |
| let SubtargetPredicate = isSICI in { |
| defm : MUBUFStore_Atomic_Pattern <BUFFER_STORE_DWORD_ADDR64, BUFFER_STORE_DWORD_OFFSET, i32, store_atomic_global>; |
| defm : MUBUFStore_Atomic_Pattern <BUFFER_STORE_DWORDX2_ADDR64, BUFFER_STORE_DWORDX2_OFFSET, i64, store_atomic_global>; |
| } // End Predicates = isSICI |
| |
| |
| multiclass MUBUFStore_Pattern <MUBUF_Pseudo Instr_OFFSET, ValueType vt, |
| PatFrag st> { |
| |
| def : GCNPat < |
| (st vt:$vdata, (MUBUFOffset v4i32:$srsrc, i32:$soffset, |
| i16:$offset, i1:$glc, i1:$slc, i1:$tfe)), |
| (Instr_OFFSET $vdata, $srsrc, $soffset, $offset, $glc, $slc, $tfe) |
| >; |
| } |
| |
| defm : MUBUFStore_Pattern <BUFFER_STORE_BYTE_OFFSET, i16, truncstorei8_global>; |
| defm : MUBUFStore_Pattern <BUFFER_STORE_SHORT_OFFSET, i16, store_global>; |
| |
| multiclass MUBUFScratchStorePat <MUBUF_Pseudo InstrOffen, |
| MUBUF_Pseudo InstrOffset, |
| ValueType vt, PatFrag st> { |
| def : GCNPat < |
| (st vt:$value, (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, |
| i32:$soffset, u16imm:$offset)), |
| (InstrOffen $value, $vaddr, $srsrc, $soffset, $offset, 0, 0, 0) |
| >; |
| |
| def : GCNPat < |
| (st vt:$value, (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, |
| u16imm:$offset)), |
| (InstrOffset $value, $srsrc, $soffset, $offset, 0, 0, 0) |
| >; |
| } |
| |
| defm : MUBUFScratchStorePat <BUFFER_STORE_BYTE_OFFEN, BUFFER_STORE_BYTE_OFFSET, i32, truncstorei8_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_SHORT_OFFEN, BUFFER_STORE_SHORT_OFFSET, i32, truncstorei16_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_BYTE_OFFEN, BUFFER_STORE_BYTE_OFFSET, i16, truncstorei8_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_SHORT_OFFEN, BUFFER_STORE_SHORT_OFFSET, i16, store_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_DWORD_OFFEN, BUFFER_STORE_DWORD_OFFSET, i32, store_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_DWORDX2_OFFEN, BUFFER_STORE_DWORDX2_OFFSET, v2i32, store_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_DWORDX4_OFFEN, BUFFER_STORE_DWORDX4_OFFSET, v4i32, store_private>; |
| |
| |
| let OtherPredicates = [D16PreservesUnusedBits] in { |
| // Hiding the extract high pattern in the PatFrag seems to not |
| // automatically increase the complexity. |
| let AddedComplexity = 1 in { |
| defm : MUBUFScratchStorePat <BUFFER_STORE_SHORT_D16_HI_OFFEN, BUFFER_STORE_SHORT_D16_HI_OFFSET, i32, store_hi16_private>; |
| defm : MUBUFScratchStorePat <BUFFER_STORE_BYTE_D16_HI_OFFEN, BUFFER_STORE_BYTE_D16_HI_OFFSET, i32, truncstorei8_hi16_private>; |
| } |
| } |
| |
| //===----------------------------------------------------------------------===// |
| // MTBUF Patterns |
| //===----------------------------------------------------------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| // tbuffer_load/store_format patterns |
| //===----------------------------------------------------------------------===// |
| |
| multiclass MTBUF_LoadIntrinsicPat<SDPatternOperator name, ValueType vt, |
| string opcode> { |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, 0, 0, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), |
| (!cast<MTBUF_Pseudo>(opcode # _OFFSET) $rsrc, $soffset, (as_i16imm $offset), |
| (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, i32:$vindex, 0, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), |
| (!cast<MTBUF_Pseudo>(opcode # _IDXEN) $vindex, $rsrc, $soffset, (as_i16imm $offset), |
| (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, 0, i32:$voffset, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), |
| (!cast<MTBUF_Pseudo>(opcode # _OFFEN) $voffset, $rsrc, $soffset, (as_i16imm $offset), |
| (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (vt (name v4i32:$rsrc, i32:$vindex, i32:$voffset, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), |
| (!cast<MTBUF_Pseudo>(opcode # _BOTHEN) |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), |
| (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| } |
| |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, i32, "TBUFFER_LOAD_FORMAT_X">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, v2i32, "TBUFFER_LOAD_FORMAT_XY">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, v4i32, "TBUFFER_LOAD_FORMAT_XYZW">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, f32, "TBUFFER_LOAD_FORMAT_X">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, v2f32, "TBUFFER_LOAD_FORMAT_XY">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load, v4f32, "TBUFFER_LOAD_FORMAT_XYZW">; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, f16, "TBUFFER_LOAD_FORMAT_D16_X_gfx80">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, v2i32, "TBUFFER_LOAD_FORMAT_D16_XY_gfx80">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, v4i32, "TBUFFER_LOAD_FORMAT_D16_XYZW_gfx80">; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, f16, "TBUFFER_LOAD_FORMAT_D16_X">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, v2f16, "TBUFFER_LOAD_FORMAT_D16_XY">; |
| defm : MTBUF_LoadIntrinsicPat<SItbuffer_load_d16, v4f16, "TBUFFER_LOAD_FORMAT_D16_XYZW">; |
| } // End HasPackedD16VMem. |
| |
| multiclass MTBUF_StoreIntrinsicPat<SDPatternOperator name, ValueType vt, |
| string opcode> { |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, 0, 0, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), |
| (!cast<MTBUF_Pseudo>(opcode # _OFFSET_exact) $vdata, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i8imm $dfmt), |
| (as_i8imm $nfmt), (as_i1imm $glc), |
| (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, i32:$vindex, 0, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), |
| (!cast<MTBUF_Pseudo>(opcode # _IDXEN_exact) $vdata, $vindex, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i8imm $dfmt), |
| (as_i8imm $nfmt), (as_i1imm $glc), |
| (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, 0, i32:$voffset, i32:$soffset, imm:$offset, |
| imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), |
| (!cast<MTBUF_Pseudo>(opcode # _OFFEN_exact) $vdata, $voffset, $rsrc, $soffset, |
| (as_i16imm $offset), (as_i8imm $dfmt), |
| (as_i8imm $nfmt), (as_i1imm $glc), |
| (as_i1imm $slc), 0) |
| >; |
| |
| def : GCNPat< |
| (name vt:$vdata, v4i32:$rsrc, i32:$vindex, i32:$voffset, i32:$soffset, |
| imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), |
| (!cast<MTBUF_Pseudo>(opcode # _BOTHEN_exact) |
| $vdata, |
| (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), |
| $rsrc, $soffset, (as_i16imm $offset), |
| (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) |
| >; |
| } |
| |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, i32, "TBUFFER_STORE_FORMAT_X">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, v2i32, "TBUFFER_STORE_FORMAT_XY">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_x3, v4i32, "TBUFFER_STORE_FORMAT_XYZ">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, v4i32, "TBUFFER_STORE_FORMAT_XYZW">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, f32, "TBUFFER_STORE_FORMAT_X">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, v2f32, "TBUFFER_STORE_FORMAT_XY">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_x3, v4f32, "TBUFFER_STORE_FORMAT_XYZ">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store, v4f32, "TBUFFER_STORE_FORMAT_XYZW">; |
| |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, f16, "TBUFFER_STORE_FORMAT_D16_X_gfx80">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, v2i32, "TBUFFER_STORE_FORMAT_D16_XY_gfx80">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, v4i32, "TBUFFER_STORE_FORMAT_D16_XYZW_gfx80">; |
| } // End HasUnpackedD16VMem. |
| |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, f16, "TBUFFER_STORE_FORMAT_D16_X">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, v2f16, "TBUFFER_STORE_FORMAT_D16_XY">; |
| defm : MTBUF_StoreIntrinsicPat<SItbuffer_store_d16, v4f16, "TBUFFER_STORE_FORMAT_D16_XYZW">; |
| } // End HasPackedD16VMem. |
| |
| //===----------------------------------------------------------------------===// |
| // Target instructions, move to the appropriate target TD file |
| //===----------------------------------------------------------------------===// |
| |
| //===----------------------------------------------------------------------===// |
| // SI |
| //===----------------------------------------------------------------------===// |
| |
| class MUBUF_Real_si <bits<7> op, MUBUF_Pseudo ps> : |
| MUBUF_Real<op, ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.SI> { |
| let AssemblerPredicate=isSICI; |
| let DecoderNamespace="SICI"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{15} = ps.addr64; |
| let Inst{16} = !if(ps.lds, 1, 0); |
| let Inst{24-18} = op; |
| let Inst{31-26} = 0x38; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{54} = !if(ps.has_slc, slc, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MUBUF_Real_AllAddr_si<bits<7> op> { |
| def _OFFSET_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _ADDR64_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_ADDR64")>; |
| def _OFFEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| multiclass MUBUF_Real_AllAddr_Lds_si<bits<7> op> { |
| |
| def _OFFSET_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET")>, |
| MUBUFLdsTable<0, NAME # "_OFFSET_si">; |
| def _ADDR64_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_ADDR64")>, |
| MUBUFLdsTable<0, NAME # "_ADDR64_si">; |
| def _OFFEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN")>, |
| MUBUFLdsTable<0, NAME # "_OFFEN_si">; |
| def _IDXEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN")>, |
| MUBUFLdsTable<0, NAME # "_IDXEN_si">; |
| def _BOTHEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN")>, |
| MUBUFLdsTable<0, NAME # "_BOTHEN_si">; |
| |
| def _LDS_OFFSET_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_OFFSET")>, |
| MUBUFLdsTable<1, NAME # "_OFFSET_si">; |
| def _LDS_ADDR64_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_ADDR64")>, |
| MUBUFLdsTable<1, NAME # "_ADDR64_si">; |
| def _LDS_OFFEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_OFFEN")>, |
| MUBUFLdsTable<1, NAME # "_OFFEN_si">; |
| def _LDS_IDXEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_IDXEN")>, |
| MUBUFLdsTable<1, NAME # "_IDXEN_si">; |
| def _LDS_BOTHEN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_BOTHEN")>, |
| MUBUFLdsTable<1, NAME # "_BOTHEN_si">; |
| } |
| |
| multiclass MUBUF_Real_Atomic_si<bits<7> op> : MUBUF_Real_AllAddr_si<op> { |
| def _OFFSET_RTN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET_RTN")>; |
| def _ADDR64_RTN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_ADDR64_RTN")>; |
| def _OFFEN_RTN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN_RTN")>; |
| def _IDXEN_RTN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN_RTN")>; |
| def _BOTHEN_RTN_si : MUBUF_Real_si <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN_RTN")>; |
| } |
| |
| defm BUFFER_LOAD_FORMAT_X : MUBUF_Real_AllAddr_Lds_si <0x00>; |
| defm BUFFER_LOAD_FORMAT_XY : MUBUF_Real_AllAddr_si <0x01>; |
| defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Real_AllAddr_si <0x02>; |
| defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Real_AllAddr_si <0x03>; |
| defm BUFFER_STORE_FORMAT_X : MUBUF_Real_AllAddr_si <0x04>; |
| defm BUFFER_STORE_FORMAT_XY : MUBUF_Real_AllAddr_si <0x05>; |
| defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Real_AllAddr_si <0x06>; |
| defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Real_AllAddr_si <0x07>; |
| defm BUFFER_LOAD_UBYTE : MUBUF_Real_AllAddr_Lds_si <0x08>; |
| defm BUFFER_LOAD_SBYTE : MUBUF_Real_AllAddr_Lds_si <0x09>; |
| defm BUFFER_LOAD_USHORT : MUBUF_Real_AllAddr_Lds_si <0x0a>; |
| defm BUFFER_LOAD_SSHORT : MUBUF_Real_AllAddr_Lds_si <0x0b>; |
| defm BUFFER_LOAD_DWORD : MUBUF_Real_AllAddr_Lds_si <0x0c>; |
| defm BUFFER_LOAD_DWORDX2 : MUBUF_Real_AllAddr_si <0x0d>; |
| defm BUFFER_LOAD_DWORDX4 : MUBUF_Real_AllAddr_si <0x0e>; |
| defm BUFFER_LOAD_DWORDX3 : MUBUF_Real_AllAddr_si <0x0f>; |
| defm BUFFER_STORE_BYTE : MUBUF_Real_AllAddr_si <0x18>; |
| defm BUFFER_STORE_SHORT : MUBUF_Real_AllAddr_si <0x1a>; |
| defm BUFFER_STORE_DWORD : MUBUF_Real_AllAddr_si <0x1c>; |
| defm BUFFER_STORE_DWORDX2 : MUBUF_Real_AllAddr_si <0x1d>; |
| defm BUFFER_STORE_DWORDX4 : MUBUF_Real_AllAddr_si <0x1e>; |
| defm BUFFER_STORE_DWORDX3 : MUBUF_Real_AllAddr_si <0x1f>; |
| |
| defm BUFFER_ATOMIC_SWAP : MUBUF_Real_Atomic_si <0x30>; |
| defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Real_Atomic_si <0x31>; |
| defm BUFFER_ATOMIC_ADD : MUBUF_Real_Atomic_si <0x32>; |
| defm BUFFER_ATOMIC_SUB : MUBUF_Real_Atomic_si <0x33>; |
| //defm BUFFER_ATOMIC_RSUB : MUBUF_Real_Atomic_si <0x34>; // isn't on CI & VI |
| defm BUFFER_ATOMIC_SMIN : MUBUF_Real_Atomic_si <0x35>; |
| defm BUFFER_ATOMIC_UMIN : MUBUF_Real_Atomic_si <0x36>; |
| defm BUFFER_ATOMIC_SMAX : MUBUF_Real_Atomic_si <0x37>; |
| defm BUFFER_ATOMIC_UMAX : MUBUF_Real_Atomic_si <0x38>; |
| defm BUFFER_ATOMIC_AND : MUBUF_Real_Atomic_si <0x39>; |
| defm BUFFER_ATOMIC_OR : MUBUF_Real_Atomic_si <0x3a>; |
| defm BUFFER_ATOMIC_XOR : MUBUF_Real_Atomic_si <0x3b>; |
| defm BUFFER_ATOMIC_INC : MUBUF_Real_Atomic_si <0x3c>; |
| defm BUFFER_ATOMIC_DEC : MUBUF_Real_Atomic_si <0x3d>; |
| |
| //defm BUFFER_ATOMIC_FCMPSWAP : MUBUF_Real_Atomic_si <0x3e>; // isn't on VI |
| //defm BUFFER_ATOMIC_FMIN : MUBUF_Real_Atomic_si <0x3f>; // isn't on VI |
| //defm BUFFER_ATOMIC_FMAX : MUBUF_Real_Atomic_si <0x40>; // isn't on VI |
| defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Real_Atomic_si <0x50>; |
| defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Real_Atomic_si <0x51>; |
| defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Real_Atomic_si <0x52>; |
| defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Real_Atomic_si <0x53>; |
| //defm BUFFER_ATOMIC_RSUB_X2 : MUBUF_Real_Atomic_si <0x54>; // isn't on CI & VI |
| defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Real_Atomic_si <0x55>; |
| defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Real_Atomic_si <0x56>; |
| defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Real_Atomic_si <0x57>; |
| defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Real_Atomic_si <0x58>; |
| defm BUFFER_ATOMIC_AND_X2 : MUBUF_Real_Atomic_si <0x59>; |
| defm BUFFER_ATOMIC_OR_X2 : MUBUF_Real_Atomic_si <0x5a>; |
| defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Real_Atomic_si <0x5b>; |
| defm BUFFER_ATOMIC_INC_X2 : MUBUF_Real_Atomic_si <0x5c>; |
| defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Real_Atomic_si <0x5d>; |
| // FIXME: Need to handle hazard for BUFFER_ATOMIC_FCMPSWAP_X2 on CI. |
| //defm BUFFER_ATOMIC_FCMPSWAP_X2 : MUBUF_Real_Atomic_si <0x5e">; // isn't on VI |
| //defm BUFFER_ATOMIC_FMIN_X2 : MUBUF_Real_Atomic_si <0x5f>; // isn't on VI |
| //defm BUFFER_ATOMIC_FMAX_X2 : MUBUF_Real_Atomic_si <0x60>; // isn't on VI |
| |
| def BUFFER_WBINVL1_SC_si : MUBUF_Real_si <0x70, BUFFER_WBINVL1_SC>; |
| def BUFFER_WBINVL1_si : MUBUF_Real_si <0x71, BUFFER_WBINVL1>; |
| |
| class MTBUF_Real_si <bits<3> op, MTBUF_Pseudo ps> : |
| MTBUF_Real<ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.SI> { |
| let AssemblerPredicate=isSICI; |
| let DecoderNamespace="SICI"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{15} = ps.addr64; |
| let Inst{18-16} = op; |
| let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); |
| let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); |
| let Inst{31-26} = 0x3a; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{54} = !if(ps.has_slc, slc, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MTBUF_Real_AllAddr_si<bits<3> op> { |
| def _OFFSET_si : MTBUF_Real_si <op, !cast<MTBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _ADDR64_si : MTBUF_Real_si <op, !cast<MTBUF_Pseudo>(NAME#"_ADDR64")>; |
| def _OFFEN_si : MTBUF_Real_si <op, !cast<MTBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_si : MTBUF_Real_si <op, !cast<MTBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_si : MTBUF_Real_si <op, !cast<MTBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| defm TBUFFER_LOAD_FORMAT_X : MTBUF_Real_AllAddr_si <0>; |
| defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Real_AllAddr_si <1>; |
| defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Real_AllAddr_si <2>; |
| defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Real_AllAddr_si <3>; |
| defm TBUFFER_STORE_FORMAT_X : MTBUF_Real_AllAddr_si <4>; |
| defm TBUFFER_STORE_FORMAT_XY : MTBUF_Real_AllAddr_si <5>; |
| defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Real_AllAddr_si <6>; |
| defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Real_AllAddr_si <7>; |
| |
| //===----------------------------------------------------------------------===// |
| // CI |
| //===----------------------------------------------------------------------===// |
| |
| class MUBUF_Real_ci <bits<7> op, MUBUF_Pseudo ps> : |
| MUBUF_Real_si<op, ps> { |
| let AssemblerPredicate=isCIOnly; |
| let DecoderNamespace="CI"; |
| } |
| |
| def BUFFER_WBINVL1_VOL_ci : MUBUF_Real_ci <0x70, BUFFER_WBINVL1_VOL>; |
| |
| |
| //===----------------------------------------------------------------------===// |
| // VI |
| //===----------------------------------------------------------------------===// |
| |
| class MUBUF_Real_vi <bits<7> op, MUBUF_Pseudo ps> : |
| MUBUF_Real<op, ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.VI> { |
| let AssemblerPredicate=isVI; |
| let DecoderNamespace="VI"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{16} = !if(ps.lds, 1, 0); |
| let Inst{17} = !if(ps.has_slc, slc, ?); |
| let Inst{24-18} = op; |
| let Inst{31-26} = 0x38; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MUBUF_Real_AllAddr_vi<bits<7> op> { |
| def _OFFSET_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _OFFEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| multiclass MUBUF_Real_AllAddr_Lds_vi<bits<7> op> { |
| |
| def _OFFSET_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET")>, |
| MUBUFLdsTable<0, NAME # "_OFFSET_vi">; |
| def _OFFEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN")>, |
| MUBUFLdsTable<0, NAME # "_OFFEN_vi">; |
| def _IDXEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN")>, |
| MUBUFLdsTable<0, NAME # "_IDXEN_vi">; |
| def _BOTHEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN")>, |
| MUBUFLdsTable<0, NAME # "_BOTHEN_vi">; |
| |
| def _LDS_OFFSET_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_OFFSET")>, |
| MUBUFLdsTable<1, NAME # "_OFFSET_vi">; |
| def _LDS_OFFEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_OFFEN")>, |
| MUBUFLdsTable<1, NAME # "_OFFEN_vi">; |
| def _LDS_IDXEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_IDXEN")>, |
| MUBUFLdsTable<1, NAME # "_IDXEN_vi">; |
| def _LDS_BOTHEN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_LDS_BOTHEN")>, |
| MUBUFLdsTable<1, NAME # "_BOTHEN_vi">; |
| } |
| |
| class MUBUF_Real_gfx80 <bits<7> op, MUBUF_Pseudo ps> : |
| MUBUF_Real<op, ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.GFX80> { |
| let AssemblerPredicate=HasUnpackedD16VMem; |
| let DecoderNamespace="GFX80_UNPACKED"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{16} = !if(ps.lds, 1, 0); |
| let Inst{17} = !if(ps.has_slc, slc, ?); |
| let Inst{24-18} = op; |
| let Inst{31-26} = 0x38; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MUBUF_Real_AllAddr_gfx80<bits<7> op> { |
| def _OFFSET_gfx80 : MUBUF_Real_gfx80 <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _OFFEN_gfx80 : MUBUF_Real_gfx80 <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_gfx80 : MUBUF_Real_gfx80 <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_gfx80 : MUBUF_Real_gfx80 <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| multiclass MUBUF_Real_Atomic_vi<bits<7> op> : |
| MUBUF_Real_AllAddr_vi<op> { |
| def _OFFSET_RTN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFSET_RTN")>; |
| def _OFFEN_RTN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_OFFEN_RTN")>; |
| def _IDXEN_RTN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_IDXEN_RTN")>; |
| def _BOTHEN_RTN_vi : MUBUF_Real_vi <op, !cast<MUBUF_Pseudo>(NAME#"_BOTHEN_RTN")>; |
| } |
| |
| defm BUFFER_LOAD_FORMAT_X : MUBUF_Real_AllAddr_Lds_vi <0x00>; |
| defm BUFFER_LOAD_FORMAT_XY : MUBUF_Real_AllAddr_vi <0x01>; |
| defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Real_AllAddr_vi <0x02>; |
| defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Real_AllAddr_vi <0x03>; |
| defm BUFFER_STORE_FORMAT_X : MUBUF_Real_AllAddr_vi <0x04>; |
| defm BUFFER_STORE_FORMAT_XY : MUBUF_Real_AllAddr_vi <0x05>; |
| defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Real_AllAddr_vi <0x06>; |
| defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Real_AllAddr_vi <0x07>; |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm BUFFER_LOAD_FORMAT_D16_X_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x08>; |
| defm BUFFER_LOAD_FORMAT_D16_XY_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x09>; |
| defm BUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0a>; |
| defm BUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0b>; |
| defm BUFFER_STORE_FORMAT_D16_X_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0c>; |
| defm BUFFER_STORE_FORMAT_D16_XY_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0d>; |
| defm BUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0e>; |
| defm BUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0f>; |
| } // End HasUnpackedD16VMem. |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm BUFFER_LOAD_FORMAT_D16_X : MUBUF_Real_AllAddr_vi <0x08>; |
| defm BUFFER_LOAD_FORMAT_D16_XY : MUBUF_Real_AllAddr_vi <0x09>; |
| defm BUFFER_LOAD_FORMAT_D16_XYZ : MUBUF_Real_AllAddr_vi <0x0a>; |
| defm BUFFER_LOAD_FORMAT_D16_XYZW : MUBUF_Real_AllAddr_vi <0x0b>; |
| defm BUFFER_STORE_FORMAT_D16_X : MUBUF_Real_AllAddr_vi <0x0c>; |
| defm BUFFER_STORE_FORMAT_D16_XY : MUBUF_Real_AllAddr_vi <0x0d>; |
| defm BUFFER_STORE_FORMAT_D16_XYZ : MUBUF_Real_AllAddr_vi <0x0e>; |
| defm BUFFER_STORE_FORMAT_D16_XYZW : MUBUF_Real_AllAddr_vi <0x0f>; |
| } // End HasPackedD16VMem. |
| defm BUFFER_LOAD_UBYTE : MUBUF_Real_AllAddr_Lds_vi <0x10>; |
| defm BUFFER_LOAD_SBYTE : MUBUF_Real_AllAddr_Lds_vi <0x11>; |
| defm BUFFER_LOAD_USHORT : MUBUF_Real_AllAddr_Lds_vi <0x12>; |
| defm BUFFER_LOAD_SSHORT : MUBUF_Real_AllAddr_Lds_vi <0x13>; |
| defm BUFFER_LOAD_DWORD : MUBUF_Real_AllAddr_Lds_vi <0x14>; |
| defm BUFFER_LOAD_DWORDX2 : MUBUF_Real_AllAddr_Lds_vi <0x15>; |
| defm BUFFER_LOAD_DWORDX3 : MUBUF_Real_AllAddr_Lds_vi <0x16>; |
| defm BUFFER_LOAD_DWORDX4 : MUBUF_Real_AllAddr_Lds_vi <0x17>; |
| defm BUFFER_STORE_BYTE : MUBUF_Real_AllAddr_vi <0x18>; |
| defm BUFFER_STORE_BYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x19>; |
| defm BUFFER_STORE_SHORT : MUBUF_Real_AllAddr_vi <0x1a>; |
| defm BUFFER_STORE_SHORT_D16_HI : MUBUF_Real_AllAddr_vi <0x1b>; |
| defm BUFFER_STORE_DWORD : MUBUF_Real_AllAddr_vi <0x1c>; |
| defm BUFFER_STORE_DWORDX2 : MUBUF_Real_AllAddr_vi <0x1d>; |
| defm BUFFER_STORE_DWORDX3 : MUBUF_Real_AllAddr_vi <0x1e>; |
| defm BUFFER_STORE_DWORDX4 : MUBUF_Real_AllAddr_vi <0x1f>; |
| |
| defm BUFFER_LOAD_UBYTE_D16 : MUBUF_Real_AllAddr_vi <0x20>; |
| defm BUFFER_LOAD_UBYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x21>; |
| defm BUFFER_LOAD_SBYTE_D16 : MUBUF_Real_AllAddr_vi <0x22>; |
| defm BUFFER_LOAD_SBYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x23>; |
| defm BUFFER_LOAD_SHORT_D16 : MUBUF_Real_AllAddr_vi <0x24>; |
| defm BUFFER_LOAD_SHORT_D16_HI : MUBUF_Real_AllAddr_vi <0x25>; |
| |
| defm BUFFER_LOAD_FORMAT_D16_HI_X : MUBUF_Real_AllAddr_vi <0x26>; |
| defm BUFFER_STORE_FORMAT_D16_HI_X : MUBUF_Real_AllAddr_vi <0x27>; |
| |
| defm BUFFER_ATOMIC_SWAP : MUBUF_Real_Atomic_vi <0x40>; |
| defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Real_Atomic_vi <0x41>; |
| defm BUFFER_ATOMIC_ADD : MUBUF_Real_Atomic_vi <0x42>; |
| defm BUFFER_ATOMIC_SUB : MUBUF_Real_Atomic_vi <0x43>; |
| defm BUFFER_ATOMIC_SMIN : MUBUF_Real_Atomic_vi <0x44>; |
| defm BUFFER_ATOMIC_UMIN : MUBUF_Real_Atomic_vi <0x45>; |
| defm BUFFER_ATOMIC_SMAX : MUBUF_Real_Atomic_vi <0x46>; |
| defm BUFFER_ATOMIC_UMAX : MUBUF_Real_Atomic_vi <0x47>; |
| defm BUFFER_ATOMIC_AND : MUBUF_Real_Atomic_vi <0x48>; |
| defm BUFFER_ATOMIC_OR : MUBUF_Real_Atomic_vi <0x49>; |
| defm BUFFER_ATOMIC_XOR : MUBUF_Real_Atomic_vi <0x4a>; |
| defm BUFFER_ATOMIC_INC : MUBUF_Real_Atomic_vi <0x4b>; |
| defm BUFFER_ATOMIC_DEC : MUBUF_Real_Atomic_vi <0x4c>; |
| |
| defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Real_Atomic_vi <0x60>; |
| defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Real_Atomic_vi <0x61>; |
| defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Real_Atomic_vi <0x62>; |
| defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Real_Atomic_vi <0x63>; |
| defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Real_Atomic_vi <0x64>; |
| defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Real_Atomic_vi <0x65>; |
| defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Real_Atomic_vi <0x66>; |
| defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Real_Atomic_vi <0x67>; |
| defm BUFFER_ATOMIC_AND_X2 : MUBUF_Real_Atomic_vi <0x68>; |
| defm BUFFER_ATOMIC_OR_X2 : MUBUF_Real_Atomic_vi <0x69>; |
| defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Real_Atomic_vi <0x6a>; |
| defm BUFFER_ATOMIC_INC_X2 : MUBUF_Real_Atomic_vi <0x6b>; |
| defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Real_Atomic_vi <0x6c>; |
| |
| def BUFFER_STORE_LDS_DWORD_vi : MUBUF_Real_vi <0x3d, BUFFER_STORE_LDS_DWORD>; |
| |
| def BUFFER_WBINVL1_vi : MUBUF_Real_vi <0x3e, BUFFER_WBINVL1>; |
| def BUFFER_WBINVL1_VOL_vi : MUBUF_Real_vi <0x3f, BUFFER_WBINVL1_VOL>; |
| |
| class MTBUF_Real_vi <bits<4> op, MTBUF_Pseudo ps> : |
| MTBUF_Real<ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.VI> { |
| let AssemblerPredicate=isVI; |
| let DecoderNamespace="VI"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{18-15} = op; |
| let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); |
| let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); |
| let Inst{31-26} = 0x3a; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{54} = !if(ps.has_slc, slc, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MTBUF_Real_AllAddr_vi<bits<4> op> { |
| def _OFFSET_vi : MTBUF_Real_vi <op, !cast<MTBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _OFFEN_vi : MTBUF_Real_vi <op, !cast<MTBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_vi : MTBUF_Real_vi <op, !cast<MTBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_vi : MTBUF_Real_vi <op, !cast<MTBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| class MTBUF_Real_gfx80 <bits<4> op, MTBUF_Pseudo ps> : |
| MTBUF_Real<ps>, |
| Enc64, |
| SIMCInstr<ps.PseudoInstr, SIEncodingFamily.GFX80> { |
| let AssemblerPredicate=HasUnpackedD16VMem; |
| let DecoderNamespace="GFX80_UNPACKED"; |
| |
| let Inst{11-0} = !if(ps.has_offset, offset, ?); |
| let Inst{12} = ps.offen; |
| let Inst{13} = ps.idxen; |
| let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); |
| let Inst{18-15} = op; |
| let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); |
| let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); |
| let Inst{31-26} = 0x3a; //encoding |
| let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); |
| let Inst{47-40} = !if(ps.has_vdata, vdata, ?); |
| let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); |
| let Inst{54} = !if(ps.has_slc, slc, ?); |
| let Inst{55} = !if(ps.has_tfe, tfe, ?); |
| let Inst{63-56} = !if(ps.has_soffset, soffset, ?); |
| } |
| |
| multiclass MTBUF_Real_AllAddr_gfx80<bits<4> op> { |
| def _OFFSET_gfx80 : MTBUF_Real_gfx80 <op, !cast<MTBUF_Pseudo>(NAME#"_OFFSET")>; |
| def _OFFEN_gfx80 : MTBUF_Real_gfx80 <op, !cast<MTBUF_Pseudo>(NAME#"_OFFEN")>; |
| def _IDXEN_gfx80 : MTBUF_Real_gfx80 <op, !cast<MTBUF_Pseudo>(NAME#"_IDXEN")>; |
| def _BOTHEN_gfx80 : MTBUF_Real_gfx80 <op, !cast<MTBUF_Pseudo>(NAME#"_BOTHEN")>; |
| } |
| |
| defm TBUFFER_LOAD_FORMAT_X : MTBUF_Real_AllAddr_vi <0x00>; |
| defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Real_AllAddr_vi <0x01>; |
| defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Real_AllAddr_vi <0x02>; |
| defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Real_AllAddr_vi <0x03>; |
| defm TBUFFER_STORE_FORMAT_X : MTBUF_Real_AllAddr_vi <0x04>; |
| defm TBUFFER_STORE_FORMAT_XY : MTBUF_Real_AllAddr_vi <0x05>; |
| defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Real_AllAddr_vi <0x06>; |
| defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Real_AllAddr_vi <0x07>; |
| let SubtargetPredicate = HasUnpackedD16VMem in { |
| defm TBUFFER_LOAD_FORMAT_D16_X_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x08>; |
| defm TBUFFER_LOAD_FORMAT_D16_XY_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x09>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0a>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0b>; |
| defm TBUFFER_STORE_FORMAT_D16_X_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0c>; |
| defm TBUFFER_STORE_FORMAT_D16_XY_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0d>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0e>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0f>; |
| } // End HasUnpackedD16VMem. |
| let SubtargetPredicate = HasPackedD16VMem in { |
| defm TBUFFER_LOAD_FORMAT_D16_X : MTBUF_Real_AllAddr_vi <0x08>; |
| defm TBUFFER_LOAD_FORMAT_D16_XY : MTBUF_Real_AllAddr_vi <0x09>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZ : MTBUF_Real_AllAddr_vi <0x0a>; |
| defm TBUFFER_LOAD_FORMAT_D16_XYZW : MTBUF_Real_AllAddr_vi <0x0b>; |
| defm TBUFFER_STORE_FORMAT_D16_X : MTBUF_Real_AllAddr_vi <0x0c>; |
| defm TBUFFER_STORE_FORMAT_D16_XY : MTBUF_Real_AllAddr_vi <0x0d>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZ : MTBUF_Real_AllAddr_vi <0x0e>; |
| defm TBUFFER_STORE_FORMAT_D16_XYZW : MTBUF_Real_AllAddr_vi <0x0f>; |
| } // End HasUnpackedD16VMem. |