//===-- BUFInstructions.td - Buffer Instruction Defintions ----------------===// // // The LLVM Compiler Infrastructure // // This file is distributed under the University of Illinois Open Source // License. See LICENSE.TXT for details. // //===----------------------------------------------------------------------===// def MUBUFAddr32 : ComplexPattern; def MUBUFAddr64 : ComplexPattern; def MUBUFAddr64Atomic : ComplexPattern; def MUBUFScratchOffen : ComplexPattern; def MUBUFScratchOffset : ComplexPattern; def MUBUFOffset : ComplexPattern; def MUBUFOffsetNoGLC : ComplexPattern; def MUBUFOffsetAtomic : ComplexPattern; def MUBUFIntrinsicOffset : ComplexPattern; def MUBUFIntrinsicVOffset : ComplexPattern; class MubufLoad : PatFrag < (ops node:$ptr), (op node:$ptr), [{ auto const AS = cast(N)->getAddressSpace(); return AS == AMDGPUASI.GLOBAL_ADDRESS || AS == AMDGPUASI.CONSTANT_ADDRESS; }]>; def mubuf_load : MubufLoad ; def mubuf_az_extloadi8 : MubufLoad ; def mubuf_sextloadi8 : MubufLoad ; def mubuf_az_extloadi16 : MubufLoad ; def mubuf_sextloadi16 : MubufLoad ; def mubuf_load_atomic : MubufLoad ; def BUFAddrKind { int Offset = 0; int OffEn = 1; int IdxEn = 2; int BothEn = 3; int Addr64 = 4; } class getAddrName { string ret = !if(!eq(addrKind, BUFAddrKind.Offset), "offset", !if(!eq(addrKind, BUFAddrKind.OffEn), "offen", !if(!eq(addrKind, BUFAddrKind.IdxEn), "idxen", !if(!eq(addrKind, BUFAddrKind.BothEn), "bothen", !if(!eq(addrKind, BUFAddrKind.Addr64), "addr64", ""))))); } class MUBUFAddr64Table { bit IsAddr64 = is_addr64; string OpName = Name; } class MUBUFLdsTable { bit IsLds = is_lds; string OpName = Name; } class MTBUFAddr64Table { bit IsAddr64 = is_addr64; string OpName = Name; } //===----------------------------------------------------------------------===// // MTBUF classes //===----------------------------------------------------------------------===// class MTBUF_Pseudo pattern=[]> : InstSI, SIMCInstr { let isPseudo = 1; let isCodeGenOnly = 1; let Size = 8; let UseNamedOperandTable = 1; string Mnemonic = opName; string AsmOperands = asmOps; let VM_CNT = 1; let EXP_CNT = 1; let MTBUF = 1; let Uses = [EXEC]; let hasSideEffects = 0; let SchedRW = [WriteVMEM]; let AsmMatchConverter = "cvtMtbuf"; bits<1> offen = 0; bits<1> idxen = 0; bits<1> addr64 = 0; bits<1> has_vdata = 1; bits<1> has_vaddr = 1; bits<1> has_glc = 1; bits<1> glc_value = 0; // the value for glc if no such operand bits<4> dfmt_value = 1; // the value for dfmt if no such operand bits<3> nfmt_value = 0; // the value for nfmt if no such operand bits<1> has_srsrc = 1; bits<1> has_soffset = 1; bits<1> has_offset = 1; bits<1> has_slc = 1; bits<1> has_tfe = 1; bits<1> has_dfmt = 1; bits<1> has_nfmt = 1; } class MTBUF_Real : InstSI { let isPseudo = 0; let isCodeGenOnly = 0; // copy relevant pseudo op flags let SubtargetPredicate = ps.SubtargetPredicate; let AsmMatchConverter = ps.AsmMatchConverter; let Constraints = ps.Constraints; let DisableEncoding = ps.DisableEncoding; let TSFlags = ps.TSFlags; bits<12> offset; bits<1> glc; bits<4> dfmt; bits<3> nfmt; bits<8> vaddr; bits<8> vdata; bits<7> srsrc; bits<1> slc; bits<1> tfe; bits<8> soffset; } class getMTBUFInsDA vdataList, list vaddrList=[]> { RegisterClass vdataClass = !if(!empty(vdataList), ?, !head(vdataList)); RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); dag InsNoData = !if(!empty(vaddrList), (ins SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe), (ins vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe) ); dag InsData = !if(!empty(vaddrList), (ins vdataClass:$vdata, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe), (ins vdataClass:$vdata, vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, DFMT:$dfmt, NFMT:$nfmt, GLC:$glc, SLC:$slc, TFE:$tfe) ); dag ret = !if(!empty(vdataList), InsNoData, InsData); } class getMTBUFIns vdataList=[]> { dag ret = !if(!eq(addrKind, BUFAddrKind.Offset), getMTBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.OffEn), getMTBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.IdxEn), getMTBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.BothEn), getMTBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.Addr64), getMTBUFInsDA.ret, (ins)))))); } class getMTBUFAsmOps { string Pfx = !if(!eq(addrKind, BUFAddrKind.Offset), "off, $srsrc, $dfmt, $nfmt, $soffset", !if(!eq(addrKind, BUFAddrKind.OffEn), "$vaddr, $srsrc, $dfmt, $nfmt, $soffset offen", !if(!eq(addrKind, BUFAddrKind.IdxEn), "$vaddr, $srsrc, $dfmt, $nfmt, $soffset idxen", !if(!eq(addrKind, BUFAddrKind.BothEn), "$vaddr, $srsrc, $dfmt, $nfmt, $soffset idxen offen", !if(!eq(addrKind, BUFAddrKind.Addr64), "$vaddr, $srsrc, $dfmt, $nfmt, $soffset addr64", ""))))); string ret = Pfx # "$offset"; } class MTBUF_SetupAddr { bits<1> offen = !if(!eq(addrKind, BUFAddrKind.OffEn), 1, !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); bits<1> idxen = !if(!eq(addrKind, BUFAddrKind.IdxEn), 1, !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); bits<1> addr64 = !if(!eq(addrKind, BUFAddrKind.Addr64), 1, 0); bits<1> has_vaddr = !if(!eq(addrKind, BUFAddrKind.Offset), 0, 1); } class MTBUF_Load_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind> : MTBUF_Pseudo.ret, " $vdata, " # getMTBUFAsmOps.ret # "$glc$slc$tfe", pattern>, MTBUF_SetupAddr { let PseudoInstr = opName # "_" # getAddrName.ret; let mayLoad = 1; let mayStore = 0; } multiclass MTBUF_Pseudo_Loads { def _OFFSET : MTBUF_Load_Pseudo , MTBUFAddr64Table<0, NAME>; def _ADDR64 : MTBUF_Load_Pseudo , MTBUFAddr64Table<1, NAME>; def _OFFEN : MTBUF_Load_Pseudo ; def _IDXEN : MTBUF_Load_Pseudo ; def _BOTHEN : MTBUF_Load_Pseudo ; let DisableWQM = 1 in { def _OFFSET_exact : MTBUF_Load_Pseudo ; def _OFFEN_exact : MTBUF_Load_Pseudo ; def _IDXEN_exact : MTBUF_Load_Pseudo ; def _BOTHEN_exact : MTBUF_Load_Pseudo ; } } class MTBUF_Store_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind, RegisterClass vdataClassCopy = vdataClass> : MTBUF_Pseudo.ret, " $vdata, " # getMTBUFAsmOps.ret # "$glc$slc$tfe", pattern>, MTBUF_SetupAddr { let PseudoInstr = opName # "_" # getAddrName.ret; let mayLoad = 0; let mayStore = 1; } multiclass MTBUF_Pseudo_Stores { def _OFFSET : MTBUF_Store_Pseudo , MTBUFAddr64Table<0, NAME>; def _ADDR64 : MTBUF_Store_Pseudo , MTBUFAddr64Table<1, NAME>; def _OFFEN : MTBUF_Store_Pseudo ; def _IDXEN : MTBUF_Store_Pseudo ; def _BOTHEN : MTBUF_Store_Pseudo ; let DisableWQM = 1 in { def _OFFSET_exact : MTBUF_Store_Pseudo ; def _OFFEN_exact : MTBUF_Store_Pseudo ; def _IDXEN_exact : MTBUF_Store_Pseudo ; def _BOTHEN_exact : MTBUF_Store_Pseudo ; } } //===----------------------------------------------------------------------===// // MUBUF classes //===----------------------------------------------------------------------===// class MUBUF_Pseudo pattern=[]> : InstSI, SIMCInstr { let isPseudo = 1; let isCodeGenOnly = 1; let Size = 8; let UseNamedOperandTable = 1; string Mnemonic = opName; string AsmOperands = asmOps; let VM_CNT = 1; let EXP_CNT = 1; let MUBUF = 1; let Uses = [EXEC]; let hasSideEffects = 0; let SchedRW = [WriteVMEM]; let AsmMatchConverter = "cvtMubuf"; bits<1> offen = 0; bits<1> idxen = 0; bits<1> addr64 = 0; bits<1> lds = 0; bits<1> has_vdata = 1; bits<1> has_vaddr = 1; bits<1> has_glc = 1; bits<1> glc_value = 0; // the value for glc if no such operand bits<1> has_srsrc = 1; bits<1> has_soffset = 1; bits<1> has_offset = 1; bits<1> has_slc = 1; bits<1> has_tfe = 1; } class MUBUF_Real op, MUBUF_Pseudo ps> : InstSI { let isPseudo = 0; let isCodeGenOnly = 0; // copy relevant pseudo op flags let SubtargetPredicate = ps.SubtargetPredicate; let AsmMatchConverter = ps.AsmMatchConverter; let Constraints = ps.Constraints; let DisableEncoding = ps.DisableEncoding; let TSFlags = ps.TSFlags; bits<12> offset; bits<1> glc; bits<8> vaddr; bits<8> vdata; bits<7> srsrc; bits<1> slc; bits<1> tfe; bits<8> soffset; } // For cache invalidation instructions. class MUBUF_Invalidate : MUBUF_Pseudo { let AsmMatchConverter = ""; let hasSideEffects = 1; let mayStore = 1; // Set everything to 0. let offen = 0; let idxen = 0; let addr64 = 0; let has_vdata = 0; let has_vaddr = 0; let has_glc = 0; let glc_value = 0; let has_srsrc = 0; let has_soffset = 0; let has_offset = 0; let has_slc = 0; let has_tfe = 0; } class getMUBUFInsDA vdataList, list vaddrList=[], bit isLds = 0> { RegisterClass vdataClass = !if(!empty(vdataList), ?, !head(vdataList)); RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); dag InsNoData = !if(!empty(vaddrList), (ins SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc), (ins vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc) ); dag InsData = !if(!empty(vaddrList), (ins vdataClass:$vdata, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc), (ins vdataClass:$vdata, vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, GLC:$glc, SLC:$slc) ); dag ret = !con( !if(!empty(vdataList), InsNoData, InsData), !if(isLds, (ins), (ins TFE:$tfe)) ); } class getMUBUFIns vdataList=[], bit isLds = 0> { dag ret = !if(!eq(addrKind, BUFAddrKind.Offset), getMUBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.OffEn), getMUBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.IdxEn), getMUBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.BothEn), getMUBUFInsDA.ret, !if(!eq(addrKind, BUFAddrKind.Addr64), getMUBUFInsDA.ret, (ins)))))); } class getMUBUFAsmOps { string Pfx = !if(!eq(addrKind, BUFAddrKind.Offset), "off, $srsrc, $soffset", !if(!eq(addrKind, BUFAddrKind.OffEn), "$vaddr, $srsrc, $soffset offen", !if(!eq(addrKind, BUFAddrKind.IdxEn), "$vaddr, $srsrc, $soffset idxen", !if(!eq(addrKind, BUFAddrKind.BothEn), "$vaddr, $srsrc, $soffset idxen offen", !if(!eq(addrKind, BUFAddrKind.Addr64), "$vaddr, $srsrc, $soffset addr64", ""))))); string ret = Pfx # "$offset"; } class MUBUF_SetupAddr { bits<1> offen = !if(!eq(addrKind, BUFAddrKind.OffEn), 1, !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); bits<1> idxen = !if(!eq(addrKind, BUFAddrKind.IdxEn), 1, !if(!eq(addrKind, BUFAddrKind.BothEn), 1 , 0)); bits<1> addr64 = !if(!eq(addrKind, BUFAddrKind.Addr64), 1, 0); bits<1> has_vaddr = !if(!eq(addrKind, BUFAddrKind.Offset), 0, 1); } class MUBUF_Load_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind> : MUBUF_Pseudo.ret, !if(HasTiedDest, (ins vdataClass:$vdata_in), (ins))), " $vdata, " # getMUBUFAsmOps.ret # "$glc$slc" # !if(isLds, " lds", "$tfe"), pattern>, MUBUF_SetupAddr { let PseudoInstr = opName # !if(isLds, "_lds", "") # "_" # getAddrName.ret; let AsmMatchConverter = !if(isLds, "cvtMubufLds", "cvtMubuf"); let Constraints = !if(HasTiedDest, "$vdata = $vdata_in", ""); let mayLoad = 1; let mayStore = 0; let maybeAtomic = 1; let Uses = !if(isLds, [EXEC, M0], [EXEC]); let has_tfe = !if(isLds, 0, 1); let lds = isLds; } // FIXME: tfe can't be an operand because it requires a separate // opcode because it needs an N+1 register class dest register. multiclass MUBUF_Pseudo_Loads { def _OFFSET : MUBUF_Load_Pseudo , MUBUFAddr64Table<0, NAME # !if(isLds, "_LDS", "")>; def _ADDR64 : MUBUF_Load_Pseudo , MUBUFAddr64Table<1, NAME # !if(isLds, "_LDS", "")>; def _OFFEN : MUBUF_Load_Pseudo ; def _IDXEN : MUBUF_Load_Pseudo ; def _BOTHEN : MUBUF_Load_Pseudo ; let DisableWQM = 1 in { def _OFFSET_exact : MUBUF_Load_Pseudo ; def _OFFEN_exact : MUBUF_Load_Pseudo ; def _IDXEN_exact : MUBUF_Load_Pseudo ; def _BOTHEN_exact : MUBUF_Load_Pseudo ; } } multiclass MUBUF_Pseudo_Loads_Lds { defm NAME : MUBUF_Pseudo_Loads; defm _LDS : MUBUF_Pseudo_Loads; } class MUBUF_Store_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind, RegisterClass vdataClassCopy = vdataClass> : MUBUF_Pseudo.ret, " $vdata, " # getMUBUFAsmOps.ret # "$glc$slc$tfe", pattern>, MUBUF_SetupAddr { let PseudoInstr = opName # "_" # getAddrName.ret; let mayLoad = 0; let mayStore = 1; let maybeAtomic = 1; } multiclass MUBUF_Pseudo_Stores { def _OFFSET : MUBUF_Store_Pseudo , MUBUFAddr64Table<0, NAME>; def _ADDR64 : MUBUF_Store_Pseudo , MUBUFAddr64Table<1, NAME>; def _OFFEN : MUBUF_Store_Pseudo ; def _IDXEN : MUBUF_Store_Pseudo ; def _BOTHEN : MUBUF_Store_Pseudo ; let DisableWQM = 1 in { def _OFFSET_exact : MUBUF_Store_Pseudo ; def _OFFEN_exact : MUBUF_Store_Pseudo ; def _IDXEN_exact : MUBUF_Store_Pseudo ; def _BOTHEN_exact : MUBUF_Store_Pseudo ; } } class MUBUF_Pseudo_Store_Lds : MUBUF_Pseudo { let mayLoad = 0; let mayStore = 1; let maybeAtomic = 1; let has_vdata = 0; let has_vaddr = 0; let has_tfe = 0; let lds = 1; let Uses = [EXEC, M0]; let AsmMatchConverter = "cvtMubufLds"; } class getMUBUFAtomicInsDA vaddrList=[]> { RegisterClass vaddrClass = !if(!empty(vaddrList), ?, !head(vaddrList)); dag ret = !if(vdata_in, !if(!empty(vaddrList), (ins vdataClass:$vdata_in, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc), (ins vdataClass:$vdata_in, vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc) ), !if(!empty(vaddrList), (ins vdataClass:$vdata, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc), (ins vdataClass:$vdata, vaddrClass:$vaddr, SReg_128:$srsrc, SCSrc_b32:$soffset, offset:$offset, SLC:$slc) )); } class getMUBUFAtomicIns { dag ret = !if(!eq(addrKind, BUFAddrKind.Offset), getMUBUFAtomicInsDA.ret, !if(!eq(addrKind, BUFAddrKind.OffEn), getMUBUFAtomicInsDA.ret, !if(!eq(addrKind, BUFAddrKind.IdxEn), getMUBUFAtomicInsDA.ret, !if(!eq(addrKind, BUFAddrKind.BothEn), getMUBUFAtomicInsDA.ret, !if(!eq(addrKind, BUFAddrKind.Addr64), getMUBUFAtomicInsDA.ret, (ins)))))); } class MUBUF_Atomic_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind> : MUBUF_Pseudo, MUBUF_SetupAddr { let mayStore = 1; let mayLoad = 1; let hasPostISelHook = 1; let hasSideEffects = 1; let DisableWQM = 1; let has_glc = 0; let has_tfe = 0; let maybeAtomic = 1; } class MUBUF_AtomicNoRet_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind, RegisterClass vdataClassCopy = vdataClass> : MUBUF_Atomic_Pseudo.ret, " $vdata, " # getMUBUFAsmOps.ret # "$slc", pattern>, AtomicNoRet.ret, 0> { let PseudoInstr = opName # "_" # getAddrName.ret; let glc_value = 0; let AsmMatchConverter = "cvtMubufAtomic"; } class MUBUF_AtomicRet_Pseudo pattern=[], // Workaround bug bz30254 int addrKindCopy = addrKind, RegisterClass vdataClassCopy = vdataClass> : MUBUF_Atomic_Pseudo.ret, " $vdata, " # getMUBUFAsmOps.ret # " glc$slc", pattern>, AtomicNoRet.ret, 1> { let PseudoInstr = opName # "_rtn_" # getAddrName.ret; let glc_value = 1; let Constraints = "$vdata = $vdata_in"; let DisableEncoding = "$vdata_in"; let AsmMatchConverter = "cvtMubufAtomicReturn"; } multiclass MUBUF_Pseudo_Atomics { def _OFFSET : MUBUF_AtomicNoRet_Pseudo , MUBUFAddr64Table <0, NAME>; def _ADDR64 : MUBUF_AtomicNoRet_Pseudo , MUBUFAddr64Table <1, NAME>; def _OFFEN : MUBUF_AtomicNoRet_Pseudo ; def _IDXEN : MUBUF_AtomicNoRet_Pseudo ; def _BOTHEN : MUBUF_AtomicNoRet_Pseudo ; def _OFFSET_RTN : MUBUF_AtomicRet_Pseudo , MUBUFAddr64Table <0, NAME # "_RTN">; def _ADDR64_RTN : MUBUF_AtomicRet_Pseudo , MUBUFAddr64Table <1, NAME # "_RTN">; def _OFFEN_RTN : MUBUF_AtomicRet_Pseudo ; def _IDXEN_RTN : MUBUF_AtomicRet_Pseudo ; def _BOTHEN_RTN : MUBUF_AtomicRet_Pseudo ; } //===----------------------------------------------------------------------===// // MUBUF Instructions //===----------------------------------------------------------------------===// defm BUFFER_LOAD_FORMAT_X : MUBUF_Pseudo_Loads_Lds < "buffer_load_format_x", VGPR_32 >; defm BUFFER_LOAD_FORMAT_XY : MUBUF_Pseudo_Loads < "buffer_load_format_xy", VReg_64 >; defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Pseudo_Loads < "buffer_load_format_xyz", VReg_96 >; defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Pseudo_Loads < "buffer_load_format_xyzw", VReg_128 >; defm BUFFER_STORE_FORMAT_X : MUBUF_Pseudo_Stores < "buffer_store_format_x", VGPR_32 >; defm BUFFER_STORE_FORMAT_XY : MUBUF_Pseudo_Stores < "buffer_store_format_xy", VReg_64 >; defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Pseudo_Stores < "buffer_store_format_xyz", VReg_96 >; defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Pseudo_Stores < "buffer_store_format_xyzw", VReg_128 >; let SubtargetPredicate = HasUnpackedD16VMem, D16Buf = 1 in { defm BUFFER_LOAD_FORMAT_D16_X_gfx80 : MUBUF_Pseudo_Loads < "buffer_load_format_d16_x", VGPR_32 >; defm BUFFER_LOAD_FORMAT_D16_XY_gfx80 : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xy", VReg_64 >; defm BUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xyz", VReg_96 >; defm BUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xyzw", VReg_128 >; defm BUFFER_STORE_FORMAT_D16_X_gfx80 : MUBUF_Pseudo_Stores < "buffer_store_format_d16_x", VGPR_32 >; defm BUFFER_STORE_FORMAT_D16_XY_gfx80 : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xy", VReg_64 >; defm BUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xyz", VReg_96 >; defm BUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xyzw", VReg_128 >; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem, D16Buf = 1 in { defm BUFFER_LOAD_FORMAT_D16_X : MUBUF_Pseudo_Loads < "buffer_load_format_d16_x", VGPR_32 >; defm BUFFER_LOAD_FORMAT_D16_XY : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xy", VGPR_32 >; defm BUFFER_LOAD_FORMAT_D16_XYZ : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xyz", VReg_64 >; defm BUFFER_LOAD_FORMAT_D16_XYZW : MUBUF_Pseudo_Loads < "buffer_load_format_d16_xyzw", VReg_64 >; defm BUFFER_STORE_FORMAT_D16_X : MUBUF_Pseudo_Stores < "buffer_store_format_d16_x", VGPR_32 >; defm BUFFER_STORE_FORMAT_D16_XY : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xy", VGPR_32 >; defm BUFFER_STORE_FORMAT_D16_XYZ : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xyz", VReg_64 >; defm BUFFER_STORE_FORMAT_D16_XYZW : MUBUF_Pseudo_Stores < "buffer_store_format_d16_xyzw", VReg_64 >; } // End HasPackedD16VMem. defm BUFFER_LOAD_UBYTE : MUBUF_Pseudo_Loads_Lds < "buffer_load_ubyte", VGPR_32, i32, mubuf_az_extloadi8 >; defm BUFFER_LOAD_SBYTE : MUBUF_Pseudo_Loads_Lds < "buffer_load_sbyte", VGPR_32, i32, mubuf_sextloadi8 >; defm BUFFER_LOAD_USHORT : MUBUF_Pseudo_Loads_Lds < "buffer_load_ushort", VGPR_32, i32, mubuf_az_extloadi16 >; defm BUFFER_LOAD_SSHORT : MUBUF_Pseudo_Loads_Lds < "buffer_load_sshort", VGPR_32, i32, mubuf_sextloadi16 >; defm BUFFER_LOAD_DWORD : MUBUF_Pseudo_Loads_Lds < "buffer_load_dword", VGPR_32, i32, mubuf_load >; defm BUFFER_LOAD_DWORDX2 : MUBUF_Pseudo_Loads < "buffer_load_dwordx2", VReg_64, v2i32, mubuf_load >; defm BUFFER_LOAD_DWORDX3 : MUBUF_Pseudo_Loads < "buffer_load_dwordx3", VReg_96, untyped, mubuf_load >; defm BUFFER_LOAD_DWORDX4 : MUBUF_Pseudo_Loads < "buffer_load_dwordx4", VReg_128, v4i32, mubuf_load >; // This is not described in AMD documentation, // but 'lds' versions of these opcodes are available // in at least GFX8+ chips. See Bug 37653. let SubtargetPredicate = isVI in { defm BUFFER_LOAD_DWORDX2_LDS : MUBUF_Pseudo_Loads < "buffer_load_dwordx2", VReg_64, v2i32, null_frag, 0, 1 >; defm BUFFER_LOAD_DWORDX3_LDS : MUBUF_Pseudo_Loads < "buffer_load_dwordx3", VReg_96, untyped, null_frag, 0, 1 >; defm BUFFER_LOAD_DWORDX4_LDS : MUBUF_Pseudo_Loads < "buffer_load_dwordx4", VReg_128, v4i32, null_frag, 0, 1 >; } defm BUFFER_STORE_BYTE : MUBUF_Pseudo_Stores < "buffer_store_byte", VGPR_32, i32, truncstorei8_global >; defm BUFFER_STORE_SHORT : MUBUF_Pseudo_Stores < "buffer_store_short", VGPR_32, i32, truncstorei16_global >; defm BUFFER_STORE_DWORD : MUBUF_Pseudo_Stores < "buffer_store_dword", VGPR_32, i32, store_global >; defm BUFFER_STORE_DWORDX2 : MUBUF_Pseudo_Stores < "buffer_store_dwordx2", VReg_64, v2i32, store_global >; defm BUFFER_STORE_DWORDX3 : MUBUF_Pseudo_Stores < "buffer_store_dwordx3", VReg_96, untyped, store_global >; defm BUFFER_STORE_DWORDX4 : MUBUF_Pseudo_Stores < "buffer_store_dwordx4", VReg_128, v4i32, store_global >; defm BUFFER_ATOMIC_SWAP : MUBUF_Pseudo_Atomics < "buffer_atomic_swap", VGPR_32, i32, atomic_swap_global >; defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Pseudo_Atomics < "buffer_atomic_cmpswap", VReg_64, v2i32, null_frag >; defm BUFFER_ATOMIC_ADD : MUBUF_Pseudo_Atomics < "buffer_atomic_add", VGPR_32, i32, atomic_add_global >; defm BUFFER_ATOMIC_SUB : MUBUF_Pseudo_Atomics < "buffer_atomic_sub", VGPR_32, i32, atomic_sub_global >; defm BUFFER_ATOMIC_SMIN : MUBUF_Pseudo_Atomics < "buffer_atomic_smin", VGPR_32, i32, atomic_min_global >; defm BUFFER_ATOMIC_UMIN : MUBUF_Pseudo_Atomics < "buffer_atomic_umin", VGPR_32, i32, atomic_umin_global >; defm BUFFER_ATOMIC_SMAX : MUBUF_Pseudo_Atomics < "buffer_atomic_smax", VGPR_32, i32, atomic_max_global >; defm BUFFER_ATOMIC_UMAX : MUBUF_Pseudo_Atomics < "buffer_atomic_umax", VGPR_32, i32, atomic_umax_global >; defm BUFFER_ATOMIC_AND : MUBUF_Pseudo_Atomics < "buffer_atomic_and", VGPR_32, i32, atomic_and_global >; defm BUFFER_ATOMIC_OR : MUBUF_Pseudo_Atomics < "buffer_atomic_or", VGPR_32, i32, atomic_or_global >; defm BUFFER_ATOMIC_XOR : MUBUF_Pseudo_Atomics < "buffer_atomic_xor", VGPR_32, i32, atomic_xor_global >; defm BUFFER_ATOMIC_INC : MUBUF_Pseudo_Atomics < "buffer_atomic_inc", VGPR_32, i32, atomic_inc_global >; defm BUFFER_ATOMIC_DEC : MUBUF_Pseudo_Atomics < "buffer_atomic_dec", VGPR_32, i32, atomic_dec_global >; defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_swap_x2", VReg_64, i64, atomic_swap_global >; defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_cmpswap_x2", VReg_128, v2i64, null_frag >; defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_add_x2", VReg_64, i64, atomic_add_global >; defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_sub_x2", VReg_64, i64, atomic_sub_global >; defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_smin_x2", VReg_64, i64, atomic_min_global >; defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_umin_x2", VReg_64, i64, atomic_umin_global >; defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_smax_x2", VReg_64, i64, atomic_max_global >; defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_umax_x2", VReg_64, i64, atomic_umax_global >; defm BUFFER_ATOMIC_AND_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_and_x2", VReg_64, i64, atomic_and_global >; defm BUFFER_ATOMIC_OR_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_or_x2", VReg_64, i64, atomic_or_global >; defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_xor_x2", VReg_64, i64, atomic_xor_global >; defm BUFFER_ATOMIC_INC_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_inc_x2", VReg_64, i64, atomic_inc_global >; defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Pseudo_Atomics < "buffer_atomic_dec_x2", VReg_64, i64, atomic_dec_global >; let SubtargetPredicate = isVI in { def BUFFER_STORE_LDS_DWORD : MUBUF_Pseudo_Store_Lds <"buffer_store_lds_dword">; } let SubtargetPredicate = isSI in { // isn't on CI & VI /* defm BUFFER_ATOMIC_RSUB : MUBUF_Pseudo_Atomics <"buffer_atomic_rsub">; defm BUFFER_ATOMIC_FCMPSWAP : MUBUF_Pseudo_Atomics <"buffer_atomic_fcmpswap">; defm BUFFER_ATOMIC_FMIN : MUBUF_Pseudo_Atomics <"buffer_atomic_fmin">; defm BUFFER_ATOMIC_FMAX : MUBUF_Pseudo_Atomics <"buffer_atomic_fmax">; defm BUFFER_ATOMIC_RSUB_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_rsub_x2">; defm BUFFER_ATOMIC_FCMPSWAP_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fcmpswap_x2">; defm BUFFER_ATOMIC_FMIN_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fmin_x2">; defm BUFFER_ATOMIC_FMAX_X2 : MUBUF_Pseudo_Atomics <"buffer_atomic_fmax_x2">; */ def BUFFER_WBINVL1_SC : MUBUF_Invalidate <"buffer_wbinvl1_sc", int_amdgcn_buffer_wbinvl1_sc>; } let SubtargetPredicate = HasD16LoadStore in { defm BUFFER_LOAD_UBYTE_D16 : MUBUF_Pseudo_Loads < "buffer_load_ubyte_d16", VGPR_32, i32, null_frag, 1 >; defm BUFFER_LOAD_UBYTE_D16_HI : MUBUF_Pseudo_Loads < "buffer_load_ubyte_d16_hi", VGPR_32, i32, null_frag, 1 >; defm BUFFER_LOAD_SBYTE_D16 : MUBUF_Pseudo_Loads < "buffer_load_sbyte_d16", VGPR_32, i32, null_frag, 1 >; defm BUFFER_LOAD_SBYTE_D16_HI : MUBUF_Pseudo_Loads < "buffer_load_sbyte_d16_hi", VGPR_32, i32, null_frag, 1 >; defm BUFFER_LOAD_SHORT_D16 : MUBUF_Pseudo_Loads < "buffer_load_short_d16", VGPR_32, i32, null_frag, 1 >; defm BUFFER_LOAD_SHORT_D16_HI : MUBUF_Pseudo_Loads < "buffer_load_short_d16_hi", VGPR_32, i32, null_frag, 1 >; defm BUFFER_STORE_BYTE_D16_HI : MUBUF_Pseudo_Stores < "buffer_store_byte_d16_hi", VGPR_32, i32 >; defm BUFFER_STORE_SHORT_D16_HI : MUBUF_Pseudo_Stores < "buffer_store_short_d16_hi", VGPR_32, i32 >; defm BUFFER_LOAD_FORMAT_D16_HI_X : MUBUF_Pseudo_Loads < "buffer_load_format_d16_hi_x", VGPR_32 >; defm BUFFER_STORE_FORMAT_D16_HI_X : MUBUF_Pseudo_Stores < "buffer_store_format_d16_hi_x", VGPR_32 >; } // End HasD16LoadStore def BUFFER_WBINVL1 : MUBUF_Invalidate <"buffer_wbinvl1", int_amdgcn_buffer_wbinvl1>; //===----------------------------------------------------------------------===// // MTBUF Instructions //===----------------------------------------------------------------------===// defm TBUFFER_LOAD_FORMAT_X : MTBUF_Pseudo_Loads <"tbuffer_load_format_x", VGPR_32>; defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Pseudo_Loads <"tbuffer_load_format_xy", VReg_64>; defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Pseudo_Loads <"tbuffer_load_format_xyz", VReg_128>; defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Pseudo_Loads <"tbuffer_load_format_xyzw", VReg_128>; defm TBUFFER_STORE_FORMAT_X : MTBUF_Pseudo_Stores <"tbuffer_store_format_x", VGPR_32>; defm TBUFFER_STORE_FORMAT_XY : MTBUF_Pseudo_Stores <"tbuffer_store_format_xy", VReg_64>; defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Pseudo_Stores <"tbuffer_store_format_xyz", VReg_128>; defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Pseudo_Stores <"tbuffer_store_format_xyzw", VReg_128>; let SubtargetPredicate = HasUnpackedD16VMem, D16Buf = 1 in { defm TBUFFER_LOAD_FORMAT_D16_X_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_x", VGPR_32>; defm TBUFFER_LOAD_FORMAT_D16_XY_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xy", VReg_64>; defm TBUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyz", VReg_96>; defm TBUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyzw", VReg_128>; defm TBUFFER_STORE_FORMAT_D16_X_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_x", VGPR_32>; defm TBUFFER_STORE_FORMAT_D16_XY_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xy", VReg_64>; defm TBUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyz", VReg_96>; defm TBUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyzw", VReg_128>; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem, D16Buf = 1 in { defm TBUFFER_LOAD_FORMAT_D16_X : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_x", VGPR_32>; defm TBUFFER_LOAD_FORMAT_D16_XY : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xy", VGPR_32>; defm TBUFFER_LOAD_FORMAT_D16_XYZ : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyz", VReg_64>; defm TBUFFER_LOAD_FORMAT_D16_XYZW : MTBUF_Pseudo_Loads <"tbuffer_load_format_d16_xyzw", VReg_64>; defm TBUFFER_STORE_FORMAT_D16_X : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_x", VGPR_32>; defm TBUFFER_STORE_FORMAT_D16_XY : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xy", VGPR_32>; defm TBUFFER_STORE_FORMAT_D16_XYZ : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyz", VReg_64>; defm TBUFFER_STORE_FORMAT_D16_XYZW : MTBUF_Pseudo_Stores <"tbuffer_store_format_d16_xyzw", VReg_64>; } // End HasPackedD16VMem. let SubtargetPredicate = isCIVI in { //===----------------------------------------------------------------------===// // Instruction definitions for CI and newer. //===----------------------------------------------------------------------===// // Remaining instructions: // BUFFER_LOAD_DWORDX3 // BUFFER_STORE_DWORDX3 def BUFFER_WBINVL1_VOL : MUBUF_Invalidate <"buffer_wbinvl1_vol", int_amdgcn_buffer_wbinvl1_vol>; } // End let SubtargetPredicate = isCIVI //===----------------------------------------------------------------------===// // MUBUF Patterns //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // buffer_load/store_format patterns //===----------------------------------------------------------------------===// multiclass MUBUF_LoadIntrinsicPat { def : GCNPat< (vt (name v4i32:$rsrc, 0, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$glc, imm:$slc)), (!cast(opcode # _OFFSET) $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$glc, imm:$slc)), (!cast(opcode # _IDXEN) $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, 0, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$glc, imm:$slc)), (!cast(opcode # _OFFEN) $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$glc, imm:$slc)), (!cast(opcode # _BOTHEN) (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; } defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; let SubtargetPredicate = HasUnpackedD16VMem in { defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; } // End HasPackedD16VMem. defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; defm : MUBUF_LoadIntrinsicPat; multiclass MUBUF_StoreIntrinsicPat { def : GCNPat< (name vt:$vdata, v4i32:$rsrc, 0, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$glc, imm:$slc), (!cast(opcode # _OFFSET_exact) $vdata, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$glc, imm:$slc), (!cast(opcode # _IDXEN_exact) $vdata, $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, 0, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$glc, imm:$slc), (!cast(opcode # _OFFEN_exact) $vdata, $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$glc, imm:$slc), (!cast(opcode # _BOTHEN_exact) $vdata, (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), 0) >; } defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; let SubtargetPredicate = HasUnpackedD16VMem in { defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; } // End HasPackedD16VMem. defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; defm : MUBUF_StoreIntrinsicPat; //===----------------------------------------------------------------------===// // buffer_atomic patterns //===----------------------------------------------------------------------===// multiclass BufferAtomicPatterns { def : GCNPat< (name i32:$vdata_in, v4i32:$rsrc, 0, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$slc), (!cast(opcode # _OFFSET_RTN) $vdata_in, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)) >; def : GCNPat< (name i32:$vdata_in, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$slc), (!cast(opcode # _IDXEN_RTN) $vdata_in, $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)) >; def : GCNPat< (name i32:$vdata_in, v4i32:$rsrc, 0, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$slc), (!cast(opcode # _OFFEN_RTN) $vdata_in, $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)) >; def : GCNPat< (name i32:$vdata_in, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$slc), (!cast(opcode # _BOTHEN_RTN) $vdata_in, (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)) >; } defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; defm : BufferAtomicPatterns; def : GCNPat< (SIbuffer_atomic_cmpswap i32:$data, i32:$cmp, v4i32:$rsrc, 0, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$slc), (EXTRACT_SUBREG (BUFFER_ATOMIC_CMPSWAP_OFFSET_RTN (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), sub0) >; def : GCNPat< (SIbuffer_atomic_cmpswap i32:$data, i32:$cmp, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicOffset i32:$soffset, i16:$offset), imm:$slc), (EXTRACT_SUBREG (BUFFER_ATOMIC_CMPSWAP_IDXEN_RTN (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), sub0) >; def : GCNPat< (SIbuffer_atomic_cmpswap i32:$data, i32:$cmp, v4i32:$rsrc, 0, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$slc), (EXTRACT_SUBREG (BUFFER_ATOMIC_CMPSWAP_OFFEN_RTN (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), sub0) >; def : GCNPat< (SIbuffer_atomic_cmpswap i32:$data, i32:$cmp, v4i32:$rsrc, i32:$vindex, (MUBUFIntrinsicVOffset i32:$soffset, i16:$offset, i32:$voffset), imm:$slc), (EXTRACT_SUBREG (BUFFER_ATOMIC_CMPSWAP_BOTHEN_RTN (REG_SEQUENCE VReg_64, $data, sub0, $cmp, sub1), (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $slc)), sub0) >; class MUBUFLoad_PatternADDR64 : GCNPat < (vt (constant_ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, i1:$glc, i1:$slc, i1:$tfe))), (Instr_ADDR64 $vaddr, $srsrc, $soffset, $offset, $glc, $slc, $tfe) >; multiclass MUBUFLoad_Atomic_Pattern { def : GCNPat < (vt (atomic_ld (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, i1:$slc))), (Instr_ADDR64 $vaddr, $srsrc, $soffset, $offset, 0, $slc, 0) >; def : GCNPat < (vt (atomic_ld (MUBUFOffsetNoGLC v4i32:$rsrc, i32:$soffset, i16:$offset))), (Instr_OFFSET $rsrc, $soffset, (as_i16imm $offset), 0, 0, 0) >; } let SubtargetPredicate = isSICI in { def : MUBUFLoad_PatternADDR64 ; def : MUBUFLoad_PatternADDR64 ; def : MUBUFLoad_PatternADDR64 ; def : MUBUFLoad_PatternADDR64 ; defm : MUBUFLoad_Atomic_Pattern ; defm : MUBUFLoad_Atomic_Pattern ; } // End SubtargetPredicate = isSICI multiclass MUBUFLoad_Pattern { def : GCNPat < (vt (ld (MUBUFOffset v4i32:$srsrc, i32:$soffset, i16:$offset, i1:$glc, i1:$slc, i1:$tfe))), (Instr_OFFSET $srsrc, $soffset, $offset, $glc, $slc, $tfe) >; } let OtherPredicates = [Has16BitInsts] in { defm : MUBUFLoad_Pattern ; defm : MUBUFLoad_Pattern ; defm : MUBUFLoad_Pattern ; defm : MUBUFLoad_Pattern ; defm : MUBUFLoad_Pattern ; } // End OtherPredicates = [Has16BitInsts] multiclass MUBUFScratchLoadPat { def : GCNPat < (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset))), (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0) >; def : GCNPat < (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))), (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0) >; } // XXX - Is it possible to have a complex pattern in a PatFrag? multiclass MUBUFScratchLoadPat_Hi16 { def : GCNPat < (build_vector vt:$lo, (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset)))), (v2i16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $lo)) >; def : GCNPat < (build_vector f16:$lo, (f16 (bitconvert (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset)))))), (v2f16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $lo)) >; def : GCNPat < (build_vector vt:$lo, (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset)))), (v2i16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $lo)) >; def : GCNPat < (build_vector f16:$lo, (f16 (bitconvert (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset)))))), (v2f16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $lo)) >; } multiclass MUBUFScratchLoadPat_Lo16 { def : GCNPat < (build_vector (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset))), (vt (Hi16Elt vt:$hi))), (v2i16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $hi)) >; def : GCNPat < (build_vector (f16 (bitconvert (vt (ld (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset))))), (f16 (Hi16Elt f16:$hi))), (v2f16 (InstrOffen $vaddr, $srsrc, $soffset, $offset, 0, 0, 0, $hi)) >; def : GCNPat < (build_vector (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))), (vt (Hi16Elt vt:$hi))), (v2i16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $hi)) >; def : GCNPat < (build_vector (f16 (bitconvert (vt (ld (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset))))), (f16 (Hi16Elt f16:$hi))), (v2f16 (InstrOffset $srsrc, $soffset, $offset, 0, 0, 0, $hi)) >; } defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; defm : MUBUFScratchLoadPat ; let OtherPredicates = [D16PreservesUnusedBits] in { defm : MUBUFScratchLoadPat_Hi16; defm : MUBUFScratchLoadPat_Hi16; defm : MUBUFScratchLoadPat_Hi16; defm : MUBUFScratchLoadPat_Lo16; defm : MUBUFScratchLoadPat_Lo16; defm : MUBUFScratchLoadPat_Lo16; } // BUFFER_LOAD_DWORD*, addr64=0 multiclass MUBUF_Load_Dword { def : GCNPat < (vt (int_SI_buffer_load_dword v4i32:$rsrc, (i32 imm), i32:$soffset, imm:$offset, 0, 0, imm:$glc, imm:$slc, imm:$tfe)), (offset $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), (as_i1imm $tfe)) >; def : GCNPat < (vt (int_SI_buffer_load_dword v4i32:$rsrc, i32:$vaddr, i32:$soffset, imm:$offset, 1, 0, imm:$glc, imm:$slc, imm:$tfe)), (offen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), (as_i1imm $tfe)) >; def : GCNPat < (vt (int_SI_buffer_load_dword v4i32:$rsrc, i32:$vaddr, i32:$soffset, imm:$offset, 0, 1, imm:$glc, imm:$slc, imm:$tfe)), (idxen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), (as_i1imm $tfe)) >; def : GCNPat < (vt (int_SI_buffer_load_dword v4i32:$rsrc, v2i32:$vaddr, i32:$soffset, imm:$offset, 1, 1, imm:$glc, imm:$slc, imm:$tfe)), (bothen $vaddr, $rsrc, $soffset, (as_i16imm $offset), (as_i1imm $glc), (as_i1imm $slc), (as_i1imm $tfe)) >; } defm : MUBUF_Load_Dword ; defm : MUBUF_Load_Dword ; defm : MUBUF_Load_Dword ; multiclass MUBUFStore_Atomic_Pattern { // Store follows atomic op convention so address is forst def : GCNPat < (atomic_st (MUBUFAddr64 v4i32:$srsrc, i64:$vaddr, i32:$soffset, i16:$offset, i1:$slc), vt:$val), (Instr_ADDR64 $val, $vaddr, $srsrc, $soffset, $offset, 0, $slc, 0) >; def : GCNPat < (atomic_st (MUBUFOffsetNoGLC v4i32:$rsrc, i32:$soffset, i16:$offset), vt:$val), (Instr_OFFSET $val, $rsrc, $soffset, (as_i16imm $offset), 0, 0, 0) >; } let SubtargetPredicate = isSICI in { defm : MUBUFStore_Atomic_Pattern ; defm : MUBUFStore_Atomic_Pattern ; } // End Predicates = isSICI multiclass MUBUFStore_Pattern { def : GCNPat < (st vt:$vdata, (MUBUFOffset v4i32:$srsrc, i32:$soffset, i16:$offset, i1:$glc, i1:$slc, i1:$tfe)), (Instr_OFFSET $vdata, $srsrc, $soffset, $offset, $glc, $slc, $tfe) >; } defm : MUBUFStore_Pattern ; defm : MUBUFStore_Pattern ; multiclass MUBUFScratchStorePat { def : GCNPat < (st vt:$value, (MUBUFScratchOffen v4i32:$srsrc, i32:$vaddr, i32:$soffset, u16imm:$offset)), (InstrOffen $value, $vaddr, $srsrc, $soffset, $offset, 0, 0, 0) >; def : GCNPat < (st vt:$value, (MUBUFScratchOffset v4i32:$srsrc, i32:$soffset, u16imm:$offset)), (InstrOffset $value, $srsrc, $soffset, $offset, 0, 0, 0) >; } defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; let OtherPredicates = [D16PreservesUnusedBits] in { // Hiding the extract high pattern in the PatFrag seems to not // automatically increase the complexity. let AddedComplexity = 1 in { defm : MUBUFScratchStorePat ; defm : MUBUFScratchStorePat ; } } //===----------------------------------------------------------------------===// // MTBUF Patterns //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // tbuffer_load/store_format patterns //===----------------------------------------------------------------------===// multiclass MTBUF_LoadIntrinsicPat { def : GCNPat< (vt (name v4i32:$rsrc, 0, 0, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), (!cast(opcode # _OFFSET) $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, i32:$vindex, 0, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), (!cast(opcode # _IDXEN) $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, 0, i32:$voffset, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), (!cast(opcode # _OFFEN) $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (vt (name v4i32:$rsrc, i32:$vindex, i32:$voffset, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc)), (!cast(opcode # _BOTHEN) (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; } defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; let SubtargetPredicate = HasUnpackedD16VMem in { defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; defm : MTBUF_LoadIntrinsicPat; } // End HasPackedD16VMem. multiclass MTBUF_StoreIntrinsicPat { def : GCNPat< (name vt:$vdata, v4i32:$rsrc, 0, 0, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), (!cast(opcode # _OFFSET_exact) $vdata, $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, i32:$vindex, 0, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), (!cast(opcode # _IDXEN_exact) $vdata, $vindex, $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, 0, i32:$voffset, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), (!cast(opcode # _OFFEN_exact) $vdata, $voffset, $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; def : GCNPat< (name vt:$vdata, v4i32:$rsrc, i32:$vindex, i32:$voffset, i32:$soffset, imm:$offset, imm:$dfmt, imm:$nfmt, imm:$glc, imm:$slc), (!cast(opcode # _BOTHEN_exact) $vdata, (REG_SEQUENCE VReg_64, $vindex, sub0, $voffset, sub1), $rsrc, $soffset, (as_i16imm $offset), (as_i8imm $dfmt), (as_i8imm $nfmt), (as_i1imm $glc), (as_i1imm $slc), 0) >; } defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; let SubtargetPredicate = HasUnpackedD16VMem in { defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; defm : MTBUF_StoreIntrinsicPat; } // End HasPackedD16VMem. //===----------------------------------------------------------------------===// // Target instructions, move to the appropriate target TD file //===----------------------------------------------------------------------===// //===----------------------------------------------------------------------===// // SI //===----------------------------------------------------------------------===// class MUBUF_Real_si op, MUBUF_Pseudo ps> : MUBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=isSICI; let DecoderNamespace="SICI"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{15} = ps.addr64; let Inst{16} = !if(ps.lds, 1, 0); let Inst{24-18} = op; let Inst{31-26} = 0x38; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{54} = !if(ps.has_slc, slc, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MUBUF_Real_AllAddr_si op> { def _OFFSET_si : MUBUF_Real_si (NAME#"_OFFSET")>; def _ADDR64_si : MUBUF_Real_si (NAME#"_ADDR64")>; def _OFFEN_si : MUBUF_Real_si (NAME#"_OFFEN")>; def _IDXEN_si : MUBUF_Real_si (NAME#"_IDXEN")>; def _BOTHEN_si : MUBUF_Real_si (NAME#"_BOTHEN")>; } multiclass MUBUF_Real_AllAddr_Lds_si op> { def _OFFSET_si : MUBUF_Real_si (NAME#"_OFFSET")>, MUBUFLdsTable<0, NAME # "_OFFSET_si">; def _ADDR64_si : MUBUF_Real_si (NAME#"_ADDR64")>, MUBUFLdsTable<0, NAME # "_ADDR64_si">; def _OFFEN_si : MUBUF_Real_si (NAME#"_OFFEN")>, MUBUFLdsTable<0, NAME # "_OFFEN_si">; def _IDXEN_si : MUBUF_Real_si (NAME#"_IDXEN")>, MUBUFLdsTable<0, NAME # "_IDXEN_si">; def _BOTHEN_si : MUBUF_Real_si (NAME#"_BOTHEN")>, MUBUFLdsTable<0, NAME # "_BOTHEN_si">; def _LDS_OFFSET_si : MUBUF_Real_si (NAME#"_LDS_OFFSET")>, MUBUFLdsTable<1, NAME # "_OFFSET_si">; def _LDS_ADDR64_si : MUBUF_Real_si (NAME#"_LDS_ADDR64")>, MUBUFLdsTable<1, NAME # "_ADDR64_si">; def _LDS_OFFEN_si : MUBUF_Real_si (NAME#"_LDS_OFFEN")>, MUBUFLdsTable<1, NAME # "_OFFEN_si">; def _LDS_IDXEN_si : MUBUF_Real_si (NAME#"_LDS_IDXEN")>, MUBUFLdsTable<1, NAME # "_IDXEN_si">; def _LDS_BOTHEN_si : MUBUF_Real_si (NAME#"_LDS_BOTHEN")>, MUBUFLdsTable<1, NAME # "_BOTHEN_si">; } multiclass MUBUF_Real_Atomic_si op> : MUBUF_Real_AllAddr_si { def _OFFSET_RTN_si : MUBUF_Real_si (NAME#"_OFFSET_RTN")>; def _ADDR64_RTN_si : MUBUF_Real_si (NAME#"_ADDR64_RTN")>; def _OFFEN_RTN_si : MUBUF_Real_si (NAME#"_OFFEN_RTN")>; def _IDXEN_RTN_si : MUBUF_Real_si (NAME#"_IDXEN_RTN")>; def _BOTHEN_RTN_si : MUBUF_Real_si (NAME#"_BOTHEN_RTN")>; } defm BUFFER_LOAD_FORMAT_X : MUBUF_Real_AllAddr_Lds_si <0x00>; defm BUFFER_LOAD_FORMAT_XY : MUBUF_Real_AllAddr_si <0x01>; defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Real_AllAddr_si <0x02>; defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Real_AllAddr_si <0x03>; defm BUFFER_STORE_FORMAT_X : MUBUF_Real_AllAddr_si <0x04>; defm BUFFER_STORE_FORMAT_XY : MUBUF_Real_AllAddr_si <0x05>; defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Real_AllAddr_si <0x06>; defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Real_AllAddr_si <0x07>; defm BUFFER_LOAD_UBYTE : MUBUF_Real_AllAddr_Lds_si <0x08>; defm BUFFER_LOAD_SBYTE : MUBUF_Real_AllAddr_Lds_si <0x09>; defm BUFFER_LOAD_USHORT : MUBUF_Real_AllAddr_Lds_si <0x0a>; defm BUFFER_LOAD_SSHORT : MUBUF_Real_AllAddr_Lds_si <0x0b>; defm BUFFER_LOAD_DWORD : MUBUF_Real_AllAddr_Lds_si <0x0c>; defm BUFFER_LOAD_DWORDX2 : MUBUF_Real_AllAddr_si <0x0d>; defm BUFFER_LOAD_DWORDX4 : MUBUF_Real_AllAddr_si <0x0e>; defm BUFFER_LOAD_DWORDX3 : MUBUF_Real_AllAddr_si <0x0f>; defm BUFFER_STORE_BYTE : MUBUF_Real_AllAddr_si <0x18>; defm BUFFER_STORE_SHORT : MUBUF_Real_AllAddr_si <0x1a>; defm BUFFER_STORE_DWORD : MUBUF_Real_AllAddr_si <0x1c>; defm BUFFER_STORE_DWORDX2 : MUBUF_Real_AllAddr_si <0x1d>; defm BUFFER_STORE_DWORDX4 : MUBUF_Real_AllAddr_si <0x1e>; defm BUFFER_STORE_DWORDX3 : MUBUF_Real_AllAddr_si <0x1f>; defm BUFFER_ATOMIC_SWAP : MUBUF_Real_Atomic_si <0x30>; defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Real_Atomic_si <0x31>; defm BUFFER_ATOMIC_ADD : MUBUF_Real_Atomic_si <0x32>; defm BUFFER_ATOMIC_SUB : MUBUF_Real_Atomic_si <0x33>; //defm BUFFER_ATOMIC_RSUB : MUBUF_Real_Atomic_si <0x34>; // isn't on CI & VI defm BUFFER_ATOMIC_SMIN : MUBUF_Real_Atomic_si <0x35>; defm BUFFER_ATOMIC_UMIN : MUBUF_Real_Atomic_si <0x36>; defm BUFFER_ATOMIC_SMAX : MUBUF_Real_Atomic_si <0x37>; defm BUFFER_ATOMIC_UMAX : MUBUF_Real_Atomic_si <0x38>; defm BUFFER_ATOMIC_AND : MUBUF_Real_Atomic_si <0x39>; defm BUFFER_ATOMIC_OR : MUBUF_Real_Atomic_si <0x3a>; defm BUFFER_ATOMIC_XOR : MUBUF_Real_Atomic_si <0x3b>; defm BUFFER_ATOMIC_INC : MUBUF_Real_Atomic_si <0x3c>; defm BUFFER_ATOMIC_DEC : MUBUF_Real_Atomic_si <0x3d>; //defm BUFFER_ATOMIC_FCMPSWAP : MUBUF_Real_Atomic_si <0x3e>; // isn't on VI //defm BUFFER_ATOMIC_FMIN : MUBUF_Real_Atomic_si <0x3f>; // isn't on VI //defm BUFFER_ATOMIC_FMAX : MUBUF_Real_Atomic_si <0x40>; // isn't on VI defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Real_Atomic_si <0x50>; defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Real_Atomic_si <0x51>; defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Real_Atomic_si <0x52>; defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Real_Atomic_si <0x53>; //defm BUFFER_ATOMIC_RSUB_X2 : MUBUF_Real_Atomic_si <0x54>; // isn't on CI & VI defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Real_Atomic_si <0x55>; defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Real_Atomic_si <0x56>; defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Real_Atomic_si <0x57>; defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Real_Atomic_si <0x58>; defm BUFFER_ATOMIC_AND_X2 : MUBUF_Real_Atomic_si <0x59>; defm BUFFER_ATOMIC_OR_X2 : MUBUF_Real_Atomic_si <0x5a>; defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Real_Atomic_si <0x5b>; defm BUFFER_ATOMIC_INC_X2 : MUBUF_Real_Atomic_si <0x5c>; defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Real_Atomic_si <0x5d>; // FIXME: Need to handle hazard for BUFFER_ATOMIC_FCMPSWAP_X2 on CI. //defm BUFFER_ATOMIC_FCMPSWAP_X2 : MUBUF_Real_Atomic_si <0x5e">; // isn't on VI //defm BUFFER_ATOMIC_FMIN_X2 : MUBUF_Real_Atomic_si <0x5f>; // isn't on VI //defm BUFFER_ATOMIC_FMAX_X2 : MUBUF_Real_Atomic_si <0x60>; // isn't on VI def BUFFER_WBINVL1_SC_si : MUBUF_Real_si <0x70, BUFFER_WBINVL1_SC>; def BUFFER_WBINVL1_si : MUBUF_Real_si <0x71, BUFFER_WBINVL1>; class MTBUF_Real_si op, MTBUF_Pseudo ps> : MTBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=isSICI; let DecoderNamespace="SICI"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{15} = ps.addr64; let Inst{18-16} = op; let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); let Inst{31-26} = 0x3a; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{54} = !if(ps.has_slc, slc, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MTBUF_Real_AllAddr_si op> { def _OFFSET_si : MTBUF_Real_si (NAME#"_OFFSET")>; def _ADDR64_si : MTBUF_Real_si (NAME#"_ADDR64")>; def _OFFEN_si : MTBUF_Real_si (NAME#"_OFFEN")>; def _IDXEN_si : MTBUF_Real_si (NAME#"_IDXEN")>; def _BOTHEN_si : MTBUF_Real_si (NAME#"_BOTHEN")>; } defm TBUFFER_LOAD_FORMAT_X : MTBUF_Real_AllAddr_si <0>; defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Real_AllAddr_si <1>; defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Real_AllAddr_si <2>; defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Real_AllAddr_si <3>; defm TBUFFER_STORE_FORMAT_X : MTBUF_Real_AllAddr_si <4>; defm TBUFFER_STORE_FORMAT_XY : MTBUF_Real_AllAddr_si <5>; defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Real_AllAddr_si <6>; defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Real_AllAddr_si <7>; //===----------------------------------------------------------------------===// // CI //===----------------------------------------------------------------------===// class MUBUF_Real_ci op, MUBUF_Pseudo ps> : MUBUF_Real_si { let AssemblerPredicate=isCIOnly; let DecoderNamespace="CI"; } def BUFFER_WBINVL1_VOL_ci : MUBUF_Real_ci <0x70, BUFFER_WBINVL1_VOL>; //===----------------------------------------------------------------------===// // VI //===----------------------------------------------------------------------===// class MUBUF_Real_vi op, MUBUF_Pseudo ps> : MUBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=isVI; let DecoderNamespace="VI"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{16} = !if(ps.lds, 1, 0); let Inst{17} = !if(ps.has_slc, slc, ?); let Inst{24-18} = op; let Inst{31-26} = 0x38; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MUBUF_Real_AllAddr_vi op> { def _OFFSET_vi : MUBUF_Real_vi (NAME#"_OFFSET")>; def _OFFEN_vi : MUBUF_Real_vi (NAME#"_OFFEN")>; def _IDXEN_vi : MUBUF_Real_vi (NAME#"_IDXEN")>; def _BOTHEN_vi : MUBUF_Real_vi (NAME#"_BOTHEN")>; } multiclass MUBUF_Real_AllAddr_Lds_vi op> { def _OFFSET_vi : MUBUF_Real_vi (NAME#"_OFFSET")>, MUBUFLdsTable<0, NAME # "_OFFSET_vi">; def _OFFEN_vi : MUBUF_Real_vi (NAME#"_OFFEN")>, MUBUFLdsTable<0, NAME # "_OFFEN_vi">; def _IDXEN_vi : MUBUF_Real_vi (NAME#"_IDXEN")>, MUBUFLdsTable<0, NAME # "_IDXEN_vi">; def _BOTHEN_vi : MUBUF_Real_vi (NAME#"_BOTHEN")>, MUBUFLdsTable<0, NAME # "_BOTHEN_vi">; def _LDS_OFFSET_vi : MUBUF_Real_vi (NAME#"_LDS_OFFSET")>, MUBUFLdsTable<1, NAME # "_OFFSET_vi">; def _LDS_OFFEN_vi : MUBUF_Real_vi (NAME#"_LDS_OFFEN")>, MUBUFLdsTable<1, NAME # "_OFFEN_vi">; def _LDS_IDXEN_vi : MUBUF_Real_vi (NAME#"_LDS_IDXEN")>, MUBUFLdsTable<1, NAME # "_IDXEN_vi">; def _LDS_BOTHEN_vi : MUBUF_Real_vi (NAME#"_LDS_BOTHEN")>, MUBUFLdsTable<1, NAME # "_BOTHEN_vi">; } class MUBUF_Real_gfx80 op, MUBUF_Pseudo ps> : MUBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=HasUnpackedD16VMem; let DecoderNamespace="GFX80_UNPACKED"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{16} = !if(ps.lds, 1, 0); let Inst{17} = !if(ps.has_slc, slc, ?); let Inst{24-18} = op; let Inst{31-26} = 0x38; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MUBUF_Real_AllAddr_gfx80 op> { def _OFFSET_gfx80 : MUBUF_Real_gfx80 (NAME#"_OFFSET")>; def _OFFEN_gfx80 : MUBUF_Real_gfx80 (NAME#"_OFFEN")>; def _IDXEN_gfx80 : MUBUF_Real_gfx80 (NAME#"_IDXEN")>; def _BOTHEN_gfx80 : MUBUF_Real_gfx80 (NAME#"_BOTHEN")>; } multiclass MUBUF_Real_Atomic_vi op> : MUBUF_Real_AllAddr_vi { def _OFFSET_RTN_vi : MUBUF_Real_vi (NAME#"_OFFSET_RTN")>; def _OFFEN_RTN_vi : MUBUF_Real_vi (NAME#"_OFFEN_RTN")>; def _IDXEN_RTN_vi : MUBUF_Real_vi (NAME#"_IDXEN_RTN")>; def _BOTHEN_RTN_vi : MUBUF_Real_vi (NAME#"_BOTHEN_RTN")>; } defm BUFFER_LOAD_FORMAT_X : MUBUF_Real_AllAddr_Lds_vi <0x00>; defm BUFFER_LOAD_FORMAT_XY : MUBUF_Real_AllAddr_vi <0x01>; defm BUFFER_LOAD_FORMAT_XYZ : MUBUF_Real_AllAddr_vi <0x02>; defm BUFFER_LOAD_FORMAT_XYZW : MUBUF_Real_AllAddr_vi <0x03>; defm BUFFER_STORE_FORMAT_X : MUBUF_Real_AllAddr_vi <0x04>; defm BUFFER_STORE_FORMAT_XY : MUBUF_Real_AllAddr_vi <0x05>; defm BUFFER_STORE_FORMAT_XYZ : MUBUF_Real_AllAddr_vi <0x06>; defm BUFFER_STORE_FORMAT_XYZW : MUBUF_Real_AllAddr_vi <0x07>; let SubtargetPredicate = HasUnpackedD16VMem in { defm BUFFER_LOAD_FORMAT_D16_X_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x08>; defm BUFFER_LOAD_FORMAT_D16_XY_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x09>; defm BUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0a>; defm BUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0b>; defm BUFFER_STORE_FORMAT_D16_X_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0c>; defm BUFFER_STORE_FORMAT_D16_XY_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0d>; defm BUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0e>; defm BUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MUBUF_Real_AllAddr_gfx80 <0x0f>; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm BUFFER_LOAD_FORMAT_D16_X : MUBUF_Real_AllAddr_vi <0x08>; defm BUFFER_LOAD_FORMAT_D16_XY : MUBUF_Real_AllAddr_vi <0x09>; defm BUFFER_LOAD_FORMAT_D16_XYZ : MUBUF_Real_AllAddr_vi <0x0a>; defm BUFFER_LOAD_FORMAT_D16_XYZW : MUBUF_Real_AllAddr_vi <0x0b>; defm BUFFER_STORE_FORMAT_D16_X : MUBUF_Real_AllAddr_vi <0x0c>; defm BUFFER_STORE_FORMAT_D16_XY : MUBUF_Real_AllAddr_vi <0x0d>; defm BUFFER_STORE_FORMAT_D16_XYZ : MUBUF_Real_AllAddr_vi <0x0e>; defm BUFFER_STORE_FORMAT_D16_XYZW : MUBUF_Real_AllAddr_vi <0x0f>; } // End HasPackedD16VMem. defm BUFFER_LOAD_UBYTE : MUBUF_Real_AllAddr_Lds_vi <0x10>; defm BUFFER_LOAD_SBYTE : MUBUF_Real_AllAddr_Lds_vi <0x11>; defm BUFFER_LOAD_USHORT : MUBUF_Real_AllAddr_Lds_vi <0x12>; defm BUFFER_LOAD_SSHORT : MUBUF_Real_AllAddr_Lds_vi <0x13>; defm BUFFER_LOAD_DWORD : MUBUF_Real_AllAddr_Lds_vi <0x14>; defm BUFFER_LOAD_DWORDX2 : MUBUF_Real_AllAddr_Lds_vi <0x15>; defm BUFFER_LOAD_DWORDX3 : MUBUF_Real_AllAddr_Lds_vi <0x16>; defm BUFFER_LOAD_DWORDX4 : MUBUF_Real_AllAddr_Lds_vi <0x17>; defm BUFFER_STORE_BYTE : MUBUF_Real_AllAddr_vi <0x18>; defm BUFFER_STORE_BYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x19>; defm BUFFER_STORE_SHORT : MUBUF_Real_AllAddr_vi <0x1a>; defm BUFFER_STORE_SHORT_D16_HI : MUBUF_Real_AllAddr_vi <0x1b>; defm BUFFER_STORE_DWORD : MUBUF_Real_AllAddr_vi <0x1c>; defm BUFFER_STORE_DWORDX2 : MUBUF_Real_AllAddr_vi <0x1d>; defm BUFFER_STORE_DWORDX3 : MUBUF_Real_AllAddr_vi <0x1e>; defm BUFFER_STORE_DWORDX4 : MUBUF_Real_AllAddr_vi <0x1f>; defm BUFFER_LOAD_UBYTE_D16 : MUBUF_Real_AllAddr_vi <0x20>; defm BUFFER_LOAD_UBYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x21>; defm BUFFER_LOAD_SBYTE_D16 : MUBUF_Real_AllAddr_vi <0x22>; defm BUFFER_LOAD_SBYTE_D16_HI : MUBUF_Real_AllAddr_vi <0x23>; defm BUFFER_LOAD_SHORT_D16 : MUBUF_Real_AllAddr_vi <0x24>; defm BUFFER_LOAD_SHORT_D16_HI : MUBUF_Real_AllAddr_vi <0x25>; defm BUFFER_LOAD_FORMAT_D16_HI_X : MUBUF_Real_AllAddr_vi <0x26>; defm BUFFER_STORE_FORMAT_D16_HI_X : MUBUF_Real_AllAddr_vi <0x27>; defm BUFFER_ATOMIC_SWAP : MUBUF_Real_Atomic_vi <0x40>; defm BUFFER_ATOMIC_CMPSWAP : MUBUF_Real_Atomic_vi <0x41>; defm BUFFER_ATOMIC_ADD : MUBUF_Real_Atomic_vi <0x42>; defm BUFFER_ATOMIC_SUB : MUBUF_Real_Atomic_vi <0x43>; defm BUFFER_ATOMIC_SMIN : MUBUF_Real_Atomic_vi <0x44>; defm BUFFER_ATOMIC_UMIN : MUBUF_Real_Atomic_vi <0x45>; defm BUFFER_ATOMIC_SMAX : MUBUF_Real_Atomic_vi <0x46>; defm BUFFER_ATOMIC_UMAX : MUBUF_Real_Atomic_vi <0x47>; defm BUFFER_ATOMIC_AND : MUBUF_Real_Atomic_vi <0x48>; defm BUFFER_ATOMIC_OR : MUBUF_Real_Atomic_vi <0x49>; defm BUFFER_ATOMIC_XOR : MUBUF_Real_Atomic_vi <0x4a>; defm BUFFER_ATOMIC_INC : MUBUF_Real_Atomic_vi <0x4b>; defm BUFFER_ATOMIC_DEC : MUBUF_Real_Atomic_vi <0x4c>; defm BUFFER_ATOMIC_SWAP_X2 : MUBUF_Real_Atomic_vi <0x60>; defm BUFFER_ATOMIC_CMPSWAP_X2 : MUBUF_Real_Atomic_vi <0x61>; defm BUFFER_ATOMIC_ADD_X2 : MUBUF_Real_Atomic_vi <0x62>; defm BUFFER_ATOMIC_SUB_X2 : MUBUF_Real_Atomic_vi <0x63>; defm BUFFER_ATOMIC_SMIN_X2 : MUBUF_Real_Atomic_vi <0x64>; defm BUFFER_ATOMIC_UMIN_X2 : MUBUF_Real_Atomic_vi <0x65>; defm BUFFER_ATOMIC_SMAX_X2 : MUBUF_Real_Atomic_vi <0x66>; defm BUFFER_ATOMIC_UMAX_X2 : MUBUF_Real_Atomic_vi <0x67>; defm BUFFER_ATOMIC_AND_X2 : MUBUF_Real_Atomic_vi <0x68>; defm BUFFER_ATOMIC_OR_X2 : MUBUF_Real_Atomic_vi <0x69>; defm BUFFER_ATOMIC_XOR_X2 : MUBUF_Real_Atomic_vi <0x6a>; defm BUFFER_ATOMIC_INC_X2 : MUBUF_Real_Atomic_vi <0x6b>; defm BUFFER_ATOMIC_DEC_X2 : MUBUF_Real_Atomic_vi <0x6c>; def BUFFER_STORE_LDS_DWORD_vi : MUBUF_Real_vi <0x3d, BUFFER_STORE_LDS_DWORD>; def BUFFER_WBINVL1_vi : MUBUF_Real_vi <0x3e, BUFFER_WBINVL1>; def BUFFER_WBINVL1_VOL_vi : MUBUF_Real_vi <0x3f, BUFFER_WBINVL1_VOL>; class MTBUF_Real_vi op, MTBUF_Pseudo ps> : MTBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=isVI; let DecoderNamespace="VI"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{18-15} = op; let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); let Inst{31-26} = 0x3a; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{54} = !if(ps.has_slc, slc, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MTBUF_Real_AllAddr_vi op> { def _OFFSET_vi : MTBUF_Real_vi (NAME#"_OFFSET")>; def _OFFEN_vi : MTBUF_Real_vi (NAME#"_OFFEN")>; def _IDXEN_vi : MTBUF_Real_vi (NAME#"_IDXEN")>; def _BOTHEN_vi : MTBUF_Real_vi (NAME#"_BOTHEN")>; } class MTBUF_Real_gfx80 op, MTBUF_Pseudo ps> : MTBUF_Real, Enc64, SIMCInstr { let AssemblerPredicate=HasUnpackedD16VMem; let DecoderNamespace="GFX80_UNPACKED"; let Inst{11-0} = !if(ps.has_offset, offset, ?); let Inst{12} = ps.offen; let Inst{13} = ps.idxen; let Inst{14} = !if(ps.has_glc, glc, ps.glc_value); let Inst{18-15} = op; let Inst{22-19} = !if(ps.has_dfmt, dfmt, ps.dfmt_value); let Inst{25-23} = !if(ps.has_nfmt, nfmt, ps.nfmt_value); let Inst{31-26} = 0x3a; //encoding let Inst{39-32} = !if(ps.has_vaddr, vaddr, ?); let Inst{47-40} = !if(ps.has_vdata, vdata, ?); let Inst{52-48} = !if(ps.has_srsrc, srsrc{6-2}, ?); let Inst{54} = !if(ps.has_slc, slc, ?); let Inst{55} = !if(ps.has_tfe, tfe, ?); let Inst{63-56} = !if(ps.has_soffset, soffset, ?); } multiclass MTBUF_Real_AllAddr_gfx80 op> { def _OFFSET_gfx80 : MTBUF_Real_gfx80 (NAME#"_OFFSET")>; def _OFFEN_gfx80 : MTBUF_Real_gfx80 (NAME#"_OFFEN")>; def _IDXEN_gfx80 : MTBUF_Real_gfx80 (NAME#"_IDXEN")>; def _BOTHEN_gfx80 : MTBUF_Real_gfx80 (NAME#"_BOTHEN")>; } defm TBUFFER_LOAD_FORMAT_X : MTBUF_Real_AllAddr_vi <0x00>; defm TBUFFER_LOAD_FORMAT_XY : MTBUF_Real_AllAddr_vi <0x01>; defm TBUFFER_LOAD_FORMAT_XYZ : MTBUF_Real_AllAddr_vi <0x02>; defm TBUFFER_LOAD_FORMAT_XYZW : MTBUF_Real_AllAddr_vi <0x03>; defm TBUFFER_STORE_FORMAT_X : MTBUF_Real_AllAddr_vi <0x04>; defm TBUFFER_STORE_FORMAT_XY : MTBUF_Real_AllAddr_vi <0x05>; defm TBUFFER_STORE_FORMAT_XYZ : MTBUF_Real_AllAddr_vi <0x06>; defm TBUFFER_STORE_FORMAT_XYZW : MTBUF_Real_AllAddr_vi <0x07>; let SubtargetPredicate = HasUnpackedD16VMem in { defm TBUFFER_LOAD_FORMAT_D16_X_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x08>; defm TBUFFER_LOAD_FORMAT_D16_XY_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x09>; defm TBUFFER_LOAD_FORMAT_D16_XYZ_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0a>; defm TBUFFER_LOAD_FORMAT_D16_XYZW_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0b>; defm TBUFFER_STORE_FORMAT_D16_X_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0c>; defm TBUFFER_STORE_FORMAT_D16_XY_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0d>; defm TBUFFER_STORE_FORMAT_D16_XYZ_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0e>; defm TBUFFER_STORE_FORMAT_D16_XYZW_gfx80 : MTBUF_Real_AllAddr_gfx80 <0x0f>; } // End HasUnpackedD16VMem. let SubtargetPredicate = HasPackedD16VMem in { defm TBUFFER_LOAD_FORMAT_D16_X : MTBUF_Real_AllAddr_vi <0x08>; defm TBUFFER_LOAD_FORMAT_D16_XY : MTBUF_Real_AllAddr_vi <0x09>; defm TBUFFER_LOAD_FORMAT_D16_XYZ : MTBUF_Real_AllAddr_vi <0x0a>; defm TBUFFER_LOAD_FORMAT_D16_XYZW : MTBUF_Real_AllAddr_vi <0x0b>; defm TBUFFER_STORE_FORMAT_D16_X : MTBUF_Real_AllAddr_vi <0x0c>; defm TBUFFER_STORE_FORMAT_D16_XY : MTBUF_Real_AllAddr_vi <0x0d>; defm TBUFFER_STORE_FORMAT_D16_XYZ : MTBUF_Real_AllAddr_vi <0x0e>; defm TBUFFER_STORE_FORMAT_D16_XYZW : MTBUF_Real_AllAddr_vi <0x0f>; } // End HasUnpackedD16VMem.