//===-- SystemZFrameLowering.cpp - Frame lowering for SystemZ -------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #include "SystemZFrameLowering.h" #include "SystemZCallingConv.h" #include "SystemZInstrBuilder.h" #include "SystemZInstrInfo.h" #include "SystemZMachineFunctionInfo.h" #include "SystemZRegisterInfo.h" #include "SystemZSubtarget.h" #include "llvm/CodeGen/LivePhysRegs.h" #include "llvm/CodeGen/MachineModuleInfo.h" #include "llvm/CodeGen/MachineRegisterInfo.h" #include "llvm/CodeGen/RegisterScavenging.h" #include "llvm/IR/Function.h" #include "llvm/Target/TargetMachine.h" using namespace llvm; namespace { // The ABI-defined register save slots, relative to the CFA (i.e. // incoming stack pointer + SystemZMC::ELFCallFrameSize). static const TargetFrameLowering::SpillSlot ELFSpillOffsetTable[] = { { SystemZ::R2D, 0x10 }, { SystemZ::R3D, 0x18 }, { SystemZ::R4D, 0x20 }, { SystemZ::R5D, 0x28 }, { SystemZ::R6D, 0x30 }, { SystemZ::R7D, 0x38 }, { SystemZ::R8D, 0x40 }, { SystemZ::R9D, 0x48 }, { SystemZ::R10D, 0x50 }, { SystemZ::R11D, 0x58 }, { SystemZ::R12D, 0x60 }, { SystemZ::R13D, 0x68 }, { SystemZ::R14D, 0x70 }, { SystemZ::R15D, 0x78 }, { SystemZ::F0D, 0x80 }, { SystemZ::F2D, 0x88 }, { SystemZ::F4D, 0x90 }, { SystemZ::F6D, 0x98 } }; static const TargetFrameLowering::SpillSlot XPLINKSpillOffsetTable[] = { {SystemZ::R4D, 0x00}, {SystemZ::R5D, 0x08}, {SystemZ::R6D, 0x10}, {SystemZ::R7D, 0x18}, {SystemZ::R8D, 0x20}, {SystemZ::R9D, 0x28}, {SystemZ::R10D, 0x30}, {SystemZ::R11D, 0x38}, {SystemZ::R12D, 0x40}, {SystemZ::R13D, 0x48}, {SystemZ::R14D, 0x50}, {SystemZ::R15D, 0x58}}; } // end anonymous namespace SystemZFrameLowering::SystemZFrameLowering(StackDirection D, Align StackAl, int LAO, Align TransAl, bool StackReal) : TargetFrameLowering(D, StackAl, LAO, TransAl, StackReal) {} std::unique_ptr SystemZFrameLowering::create(const SystemZSubtarget &STI) { if (STI.isTargetXPLINK64()) return std::make_unique(); return std::make_unique(); } MachineBasicBlock::iterator SystemZFrameLowering::eliminateCallFramePseudoInstr( MachineFunction &MF, MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const { switch (MI->getOpcode()) { case SystemZ::ADJCALLSTACKDOWN: case SystemZ::ADJCALLSTACKUP: assert(hasReservedCallFrame(MF) && "ADJSTACKDOWN and ADJSTACKUP should be no-ops"); return MBB.erase(MI); break; default: llvm_unreachable("Unexpected call frame instruction"); } } namespace { struct SZFrameSortingObj { bool IsValid = false; // True if we care about this Object. uint32_t ObjectIndex = 0; // Index of Object into MFI list. uint64_t ObjectSize = 0; // Size of Object in bytes. uint32_t D12Count = 0; // 12-bit displacement only. uint32_t DPairCount = 0; // 12 or 20 bit displacement. }; typedef std::vector SZFrameObjVec; } // namespace // TODO: Move to base class. void SystemZELFFrameLowering::orderFrameObjects( const MachineFunction &MF, SmallVectorImpl &ObjectsToAllocate) const { const MachineFrameInfo &MFI = MF.getFrameInfo(); auto *TII = MF.getSubtarget().getInstrInfo(); // Make a vector of sorting objects to track all MFI objects and mark those // to be sorted as valid. if (ObjectsToAllocate.size() <= 1) return; SZFrameObjVec SortingObjects(MFI.getObjectIndexEnd()); for (auto &Obj : ObjectsToAllocate) { SortingObjects[Obj].IsValid = true; SortingObjects[Obj].ObjectIndex = Obj; SortingObjects[Obj].ObjectSize = MFI.getObjectSize(Obj); } // Examine uses for each object and record short (12-bit) and "pair" // displacement types. for (auto &MBB : MF) for (auto &MI : MBB) { if (MI.isDebugInstr()) continue; for (unsigned I = 0, E = MI.getNumOperands(); I != E; ++I) { const MachineOperand &MO = MI.getOperand(I); if (!MO.isFI()) continue; int Index = MO.getIndex(); if (Index >= 0 && Index < MFI.getObjectIndexEnd() && SortingObjects[Index].IsValid) { if (TII->hasDisplacementPairInsn(MI.getOpcode())) SortingObjects[Index].DPairCount++; else if (!(MI.getDesc().TSFlags & SystemZII::Has20BitOffset)) SortingObjects[Index].D12Count++; } } } // Sort all objects for short/paired displacements, which should be // sufficient as it seems like all frame objects typically are within the // long displacement range. Sorting works by computing the "density" as // Count / ObjectSize. The comparisons of two such fractions are refactored // by multiplying both sides with A.ObjectSize * B.ObjectSize, in order to // eliminate the (fp) divisions. A higher density object needs to go after // in the list in order for it to end up lower on the stack. auto CmpD12 = [](const SZFrameSortingObj &A, const SZFrameSortingObj &B) { // Put all invalid and variable sized objects at the end. if (!A.IsValid || !B.IsValid) return A.IsValid; if (!A.ObjectSize || !B.ObjectSize) return A.ObjectSize > 0; uint64_t ADensityCmp = A.D12Count * B.ObjectSize; uint64_t BDensityCmp = B.D12Count * A.ObjectSize; if (ADensityCmp != BDensityCmp) return ADensityCmp < BDensityCmp; return A.DPairCount * B.ObjectSize < B.DPairCount * A.ObjectSize; }; std::stable_sort(SortingObjects.begin(), SortingObjects.end(), CmpD12); // Now modify the original list to represent the final order that // we want. unsigned Idx = 0; for (auto &Obj : SortingObjects) { // All invalid items are sorted at the end, so it's safe to stop. if (!Obj.IsValid) break; ObjectsToAllocate[Idx++] = Obj.ObjectIndex; } } bool SystemZFrameLowering::hasReservedCallFrame( const MachineFunction &MF) const { // The ELF ABI requires us to allocate 160 bytes of stack space for the // callee, with any outgoing stack arguments being placed above that. It // seems better to make that area a permanent feature of the frame even if // we're using a frame pointer. Similarly, 64-bit XPLINK requires 96 bytes // of stack space for the register save area. return true; } bool SystemZELFFrameLowering::assignCalleeSavedSpillSlots( MachineFunction &MF, const TargetRegisterInfo *TRI, std::vector &CSI) const { SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineFrameInfo &MFFrame = MF.getFrameInfo(); bool IsVarArg = MF.getFunction().isVarArg(); if (CSI.empty()) return true; // Early exit if no callee saved registers are modified! unsigned LowGPR = 0; unsigned HighGPR = SystemZ::R15D; int StartSPOffset = SystemZMC::ELFCallFrameSize; for (auto &CS : CSI) { Register Reg = CS.getReg(); int Offset = getRegSpillOffset(MF, Reg); if (Offset) { if (SystemZ::GR64BitRegClass.contains(Reg) && StartSPOffset > Offset) { LowGPR = Reg; StartSPOffset = Offset; } Offset -= SystemZMC::ELFCallFrameSize; int FrameIdx = MFFrame.CreateFixedSpillStackObject(8, Offset); CS.setFrameIdx(FrameIdx); } else CS.setFrameIdx(INT32_MAX); } // Save the range of call-saved registers, for use by the // prologue/epilogue inserters. ZFI->setRestoreGPRRegs(LowGPR, HighGPR, StartSPOffset); if (IsVarArg) { // Also save the GPR varargs, if any. R6D is call-saved, so would // already be included, but we also need to handle the call-clobbered // argument registers. Register FirstGPR = ZFI->getVarArgsFirstGPR(); if (FirstGPR < SystemZ::ELFNumArgGPRs) { unsigned Reg = SystemZ::ELFArgGPRs[FirstGPR]; int Offset = getRegSpillOffset(MF, Reg); if (StartSPOffset > Offset) { LowGPR = Reg; StartSPOffset = Offset; } } } ZFI->setSpillGPRRegs(LowGPR, HighGPR, StartSPOffset); // Create fixed stack objects for the remaining registers. int CurrOffset = -SystemZMC::ELFCallFrameSize; if (usePackedStack(MF)) CurrOffset += StartSPOffset; for (auto &CS : CSI) { if (CS.getFrameIdx() != INT32_MAX) continue; Register Reg = CS.getReg(); const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(Reg); unsigned Size = TRI->getSpillSize(*RC); CurrOffset -= Size; assert(CurrOffset % 8 == 0 && "8-byte alignment required for for all register save slots"); int FrameIdx = MFFrame.CreateFixedSpillStackObject(Size, CurrOffset); CS.setFrameIdx(FrameIdx); } return true; } void SystemZELFFrameLowering::determineCalleeSaves(MachineFunction &MF, BitVector &SavedRegs, RegScavenger *RS) const { TargetFrameLowering::determineCalleeSaves(MF, SavedRegs, RS); MachineFrameInfo &MFFrame = MF.getFrameInfo(); const TargetRegisterInfo *TRI = MF.getSubtarget().getRegisterInfo(); bool HasFP = hasFP(MF); SystemZMachineFunctionInfo *MFI = MF.getInfo(); bool IsVarArg = MF.getFunction().isVarArg(); // va_start stores incoming FPR varargs in the normal way, but delegates // the saving of incoming GPR varargs to spillCalleeSavedRegisters(). // Record these pending uses, which typically include the call-saved // argument register R6D. if (IsVarArg) for (unsigned I = MFI->getVarArgsFirstGPR(); I < SystemZ::ELFNumArgGPRs; ++I) SavedRegs.set(SystemZ::ELFArgGPRs[I]); // If there are any landing pads, entering them will modify r6/r7. if (!MF.getLandingPads().empty()) { SavedRegs.set(SystemZ::R6D); SavedRegs.set(SystemZ::R7D); } // If the function requires a frame pointer, record that the hard // frame pointer will be clobbered. if (HasFP) SavedRegs.set(SystemZ::R11D); // If the function calls other functions, record that the return // address register will be clobbered. if (MFFrame.hasCalls()) SavedRegs.set(SystemZ::R14D); // If we are saving GPRs other than the stack pointer, we might as well // save and restore the stack pointer at the same time, via STMG and LMG. // This allows the deallocation to be done by the LMG, rather than needing // a separate %r15 addition. const MCPhysReg *CSRegs = TRI->getCalleeSavedRegs(&MF); for (unsigned I = 0; CSRegs[I]; ++I) { unsigned Reg = CSRegs[I]; if (SystemZ::GR64BitRegClass.contains(Reg) && SavedRegs.test(Reg)) { SavedRegs.set(SystemZ::R15D); break; } } } SystemZELFFrameLowering::SystemZELFFrameLowering() : SystemZFrameLowering(TargetFrameLowering::StackGrowsDown, Align(8), 0, Align(8), /* StackRealignable */ false), RegSpillOffsets(0) { // Due to the SystemZ ABI, the DWARF CFA (Canonical Frame Address) is not // equal to the incoming stack pointer, but to incoming stack pointer plus // 160. Instead of using a Local Area Offset, the Register save area will // be occupied by fixed frame objects, and all offsets are actually // relative to CFA. // Create a mapping from register number to save slot offset. // These offsets are relative to the start of the register save area. RegSpillOffsets.grow(SystemZ::NUM_TARGET_REGS); for (unsigned I = 0, E = array_lengthof(ELFSpillOffsetTable); I != E; ++I) RegSpillOffsets[ELFSpillOffsetTable[I].Reg] = ELFSpillOffsetTable[I].Offset; } // Add GPR64 to the save instruction being built by MIB, which is in basic // block MBB. IsImplicit says whether this is an explicit operand to the // instruction, or an implicit one that comes between the explicit start // and end registers. static void addSavedGPR(MachineBasicBlock &MBB, MachineInstrBuilder &MIB, unsigned GPR64, bool IsImplicit) { const TargetRegisterInfo *RI = MBB.getParent()->getSubtarget().getRegisterInfo(); Register GPR32 = RI->getSubReg(GPR64, SystemZ::subreg_l32); bool IsLive = MBB.isLiveIn(GPR64) || MBB.isLiveIn(GPR32); if (!IsLive || !IsImplicit) { MIB.addReg(GPR64, getImplRegState(IsImplicit) | getKillRegState(!IsLive)); if (!IsLive) MBB.addLiveIn(GPR64); } } bool SystemZELFFrameLowering::spillCalleeSavedRegisters( MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, ArrayRef CSI, const TargetRegisterInfo *TRI) const { if (CSI.empty()) return false; MachineFunction &MF = *MBB.getParent(); const TargetInstrInfo *TII = MF.getSubtarget().getInstrInfo(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); bool IsVarArg = MF.getFunction().isVarArg(); DebugLoc DL; // Save GPRs SystemZ::GPRRegs SpillGPRs = ZFI->getSpillGPRRegs(); if (SpillGPRs.LowGPR) { assert(SpillGPRs.LowGPR != SpillGPRs.HighGPR && "Should be saving %r15 and something else"); // Build an STMG instruction. MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::STMG)); // Add the explicit register operands. addSavedGPR(MBB, MIB, SpillGPRs.LowGPR, false); addSavedGPR(MBB, MIB, SpillGPRs.HighGPR, false); // Add the address. MIB.addReg(SystemZ::R15D).addImm(SpillGPRs.GPROffset); // Make sure all call-saved GPRs are included as operands and are // marked as live on entry. for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (SystemZ::GR64BitRegClass.contains(Reg)) addSavedGPR(MBB, MIB, Reg, true); } // ...likewise GPR varargs. if (IsVarArg) for (unsigned I = ZFI->getVarArgsFirstGPR(); I < SystemZ::ELFNumArgGPRs; ++I) addSavedGPR(MBB, MIB, SystemZ::ELFArgGPRs[I], true); } // Save FPRs/VRs in the normal TargetInstrInfo way. for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (SystemZ::FP64BitRegClass.contains(Reg)) { MBB.addLiveIn(Reg); TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(), &SystemZ::FP64BitRegClass, TRI); } if (SystemZ::VR128BitRegClass.contains(Reg)) { MBB.addLiveIn(Reg); TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(), &SystemZ::VR128BitRegClass, TRI); } } return true; } bool SystemZELFFrameLowering::restoreCalleeSavedRegisters( MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, MutableArrayRef CSI, const TargetRegisterInfo *TRI) const { if (CSI.empty()) return false; MachineFunction &MF = *MBB.getParent(); const TargetInstrInfo *TII = MF.getSubtarget().getInstrInfo(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); bool HasFP = hasFP(MF); DebugLoc DL = MBBI != MBB.end() ? MBBI->getDebugLoc() : DebugLoc(); // Restore FPRs/VRs in the normal TargetInstrInfo way. for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (SystemZ::FP64BitRegClass.contains(Reg)) TII->loadRegFromStackSlot(MBB, MBBI, Reg, I.getFrameIdx(), &SystemZ::FP64BitRegClass, TRI); if (SystemZ::VR128BitRegClass.contains(Reg)) TII->loadRegFromStackSlot(MBB, MBBI, Reg, I.getFrameIdx(), &SystemZ::VR128BitRegClass, TRI); } // Restore call-saved GPRs (but not call-clobbered varargs, which at // this point might hold return values). SystemZ::GPRRegs RestoreGPRs = ZFI->getRestoreGPRRegs(); if (RestoreGPRs.LowGPR) { // If we saved any of %r2-%r5 as varargs, we should also be saving // and restoring %r6. If we're saving %r6 or above, we should be // restoring it too. assert(RestoreGPRs.LowGPR != RestoreGPRs.HighGPR && "Should be loading %r15 and something else"); // Build an LMG instruction. MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LMG)); // Add the explicit register operands. MIB.addReg(RestoreGPRs.LowGPR, RegState::Define); MIB.addReg(RestoreGPRs.HighGPR, RegState::Define); // Add the address. MIB.addReg(HasFP ? SystemZ::R11D : SystemZ::R15D); MIB.addImm(RestoreGPRs.GPROffset); // Do a second scan adding regs as being defined by instruction for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (Reg != RestoreGPRs.LowGPR && Reg != RestoreGPRs.HighGPR && SystemZ::GR64BitRegClass.contains(Reg)) MIB.addReg(Reg, RegState::ImplicitDefine); } } return true; } void SystemZELFFrameLowering::processFunctionBeforeFrameFinalized( MachineFunction &MF, RegScavenger *RS) const { MachineFrameInfo &MFFrame = MF.getFrameInfo(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineRegisterInfo *MRI = &MF.getRegInfo(); bool BackChain = MF.getFunction().hasFnAttribute("backchain"); if (!usePackedStack(MF) || BackChain) // Create the incoming register save area. getOrCreateFramePointerSaveIndex(MF); // Get the size of our stack frame to be allocated ... uint64_t StackSize = (MFFrame.estimateStackSize(MF) + SystemZMC::ELFCallFrameSize); // ... and the maximum offset we may need to reach into the // caller's frame to access the save area or stack arguments. int64_t MaxArgOffset = 0; for (int I = MFFrame.getObjectIndexBegin(); I != 0; ++I) if (MFFrame.getObjectOffset(I) >= 0) { int64_t ArgOffset = MFFrame.getObjectOffset(I) + MFFrame.getObjectSize(I); MaxArgOffset = std::max(MaxArgOffset, ArgOffset); } uint64_t MaxReach = StackSize + MaxArgOffset; if (!isUInt<12>(MaxReach)) { // We may need register scavenging slots if some parts of the frame // are outside the reach of an unsigned 12-bit displacement. // Create 2 for the case where both addresses in an MVC are // out of range. RS->addScavengingFrameIndex(MFFrame.CreateStackObject(8, Align(8), false)); RS->addScavengingFrameIndex(MFFrame.CreateStackObject(8, Align(8), false)); } // If R6 is used as an argument register it is still callee saved. If it in // this case is not clobbered (and restored) it should never be marked as // killed. if (MF.front().isLiveIn(SystemZ::R6D) && ZFI->getRestoreGPRRegs().LowGPR != SystemZ::R6D) for (auto &MO : MRI->use_nodbg_operands(SystemZ::R6D)) MO.setIsKill(false); } // Emit instructions before MBBI (in MBB) to add NumBytes to Reg. static void emitIncrement(MachineBasicBlock &MBB, MachineBasicBlock::iterator &MBBI, const DebugLoc &DL, Register Reg, int64_t NumBytes, const TargetInstrInfo *TII) { while (NumBytes) { unsigned Opcode; int64_t ThisVal = NumBytes; if (isInt<16>(NumBytes)) Opcode = SystemZ::AGHI; else { Opcode = SystemZ::AGFI; // Make sure we maintain 8-byte stack alignment. int64_t MinVal = -uint64_t(1) << 31; int64_t MaxVal = (int64_t(1) << 31) - 8; if (ThisVal < MinVal) ThisVal = MinVal; else if (ThisVal > MaxVal) ThisVal = MaxVal; } MachineInstr *MI = BuildMI(MBB, MBBI, DL, TII->get(Opcode), Reg) .addReg(Reg).addImm(ThisVal); // The CC implicit def is dead. MI->getOperand(3).setIsDead(); NumBytes -= ThisVal; } } // Add CFI for the new CFA offset. static void buildCFAOffs(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, int Offset, const SystemZInstrInfo *ZII) { unsigned CFIIndex = MBB.getParent()->addFrameInst( MCCFIInstruction::cfiDefCfaOffset(nullptr, -Offset)); BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION)) .addCFIIndex(CFIIndex); } // Add CFI for the new frame location. static void buildDefCFAReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned Reg, const SystemZInstrInfo *ZII) { MachineFunction &MF = *MBB.getParent(); MachineModuleInfo &MMI = MF.getMMI(); const MCRegisterInfo *MRI = MMI.getContext().getRegisterInfo(); unsigned RegNum = MRI->getDwarfRegNum(Reg, true); unsigned CFIIndex = MF.addFrameInst( MCCFIInstruction::createDefCfaRegister(nullptr, RegNum)); BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION)) .addCFIIndex(CFIIndex); } void SystemZELFFrameLowering::emitPrologue(MachineFunction &MF, MachineBasicBlock &MBB) const { assert(&MF.front() == &MBB && "Shrink-wrapping not yet supported"); const SystemZSubtarget &STI = MF.getSubtarget(); const SystemZTargetLowering &TLI = *STI.getTargetLowering(); MachineFrameInfo &MFFrame = MF.getFrameInfo(); auto *ZII = static_cast(STI.getInstrInfo()); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineBasicBlock::iterator MBBI = MBB.begin(); MachineModuleInfo &MMI = MF.getMMI(); const MCRegisterInfo *MRI = MMI.getContext().getRegisterInfo(); const std::vector &CSI = MFFrame.getCalleeSavedInfo(); bool HasFP = hasFP(MF); // In GHC calling convention C stack space, including the ABI-defined // 160-byte base area, is (de)allocated by GHC itself. This stack space may // be used by LLVM as spill slots for the tail recursive GHC functions. Thus // do not allocate stack space here, too. if (MF.getFunction().getCallingConv() == CallingConv::GHC) { if (MFFrame.getStackSize() > 2048 * sizeof(long)) { report_fatal_error( "Pre allocated stack space for GHC function is too small"); } if (HasFP) { report_fatal_error( "In GHC calling convention a frame pointer is not supported"); } MFFrame.setStackSize(MFFrame.getStackSize() + SystemZMC::ELFCallFrameSize); return; } // Debug location must be unknown since the first debug location is used // to determine the end of the prologue. DebugLoc DL; // The current offset of the stack pointer from the CFA. int64_t SPOffsetFromCFA = -SystemZMC::ELFCFAOffsetFromInitialSP; if (ZFI->getSpillGPRRegs().LowGPR) { // Skip over the GPR saves. if (MBBI != MBB.end() && MBBI->getOpcode() == SystemZ::STMG) ++MBBI; else llvm_unreachable("Couldn't skip over GPR saves"); // Add CFI for the GPR saves. for (auto &Save : CSI) { Register Reg = Save.getReg(); if (SystemZ::GR64BitRegClass.contains(Reg)) { int FI = Save.getFrameIdx(); int64_t Offset = MFFrame.getObjectOffset(FI); unsigned CFIIndex = MF.addFrameInst(MCCFIInstruction::createOffset( nullptr, MRI->getDwarfRegNum(Reg, true), Offset)); BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION)) .addCFIIndex(CFIIndex); } } } uint64_t StackSize = MFFrame.getStackSize(); // We need to allocate the ABI-defined 160-byte base area whenever // we allocate stack space for our own use and whenever we call another // function. bool HasStackObject = false; for (unsigned i = 0, e = MFFrame.getObjectIndexEnd(); i != e; ++i) if (!MFFrame.isDeadObjectIndex(i)) { HasStackObject = true; break; } if (HasStackObject || MFFrame.hasCalls()) StackSize += SystemZMC::ELFCallFrameSize; // Don't allocate the incoming reg save area. StackSize = StackSize > SystemZMC::ELFCallFrameSize ? StackSize - SystemZMC::ELFCallFrameSize : 0; MFFrame.setStackSize(StackSize); if (StackSize) { // Allocate StackSize bytes. int64_t Delta = -int64_t(StackSize); const unsigned ProbeSize = TLI.getStackProbeSize(MF); bool FreeProbe = (ZFI->getSpillGPRRegs().GPROffset && (ZFI->getSpillGPRRegs().GPROffset + StackSize) < ProbeSize); if (!FreeProbe && MF.getSubtarget().getTargetLowering()->hasInlineStackProbe(MF)) { // Stack probing may involve looping, but splitting the prologue block // is not possible at this point since it would invalidate the // SaveBlocks / RestoreBlocks sets of PEI in the single block function // case. Build a pseudo to be handled later by inlineStackProbe(). BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::PROBED_STACKALLOC)) .addImm(StackSize); } else { bool StoreBackchain = MF.getFunction().hasFnAttribute("backchain"); // If we need backchain, save current stack pointer. R1 is free at // this point. if (StoreBackchain) BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR)) .addReg(SystemZ::R1D, RegState::Define).addReg(SystemZ::R15D); emitIncrement(MBB, MBBI, DL, SystemZ::R15D, Delta, ZII); buildCFAOffs(MBB, MBBI, DL, SPOffsetFromCFA + Delta, ZII); if (StoreBackchain) BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::STG)) .addReg(SystemZ::R1D, RegState::Kill).addReg(SystemZ::R15D) .addImm(getBackchainOffset(MF)).addReg(0); } SPOffsetFromCFA += Delta; } if (HasFP) { // Copy the base of the frame to R11. BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR), SystemZ::R11D) .addReg(SystemZ::R15D); // Add CFI for the new frame location. buildDefCFAReg(MBB, MBBI, DL, SystemZ::R11D, ZII); // Mark the FramePtr as live at the beginning of every block except // the entry block. (We'll have marked R11 as live on entry when // saving the GPRs.) for (MachineBasicBlock &MBBJ : llvm::drop_begin(MF)) MBBJ.addLiveIn(SystemZ::R11D); } // Skip over the FPR/VR saves. SmallVector CFIIndexes; for (auto &Save : CSI) { Register Reg = Save.getReg(); if (SystemZ::FP64BitRegClass.contains(Reg)) { if (MBBI != MBB.end() && (MBBI->getOpcode() == SystemZ::STD || MBBI->getOpcode() == SystemZ::STDY)) ++MBBI; else llvm_unreachable("Couldn't skip over FPR save"); } else if (SystemZ::VR128BitRegClass.contains(Reg)) { if (MBBI != MBB.end() && MBBI->getOpcode() == SystemZ::VST) ++MBBI; else llvm_unreachable("Couldn't skip over VR save"); } else continue; // Add CFI for the this save. unsigned DwarfReg = MRI->getDwarfRegNum(Reg, true); Register IgnoredFrameReg; int64_t Offset = getFrameIndexReference(MF, Save.getFrameIdx(), IgnoredFrameReg) .getFixed(); unsigned CFIIndex = MF.addFrameInst(MCCFIInstruction::createOffset( nullptr, DwarfReg, SPOffsetFromCFA + Offset)); CFIIndexes.push_back(CFIIndex); } // Complete the CFI for the FPR/VR saves, modelling them as taking effect // after the last save. for (auto CFIIndex : CFIIndexes) { BuildMI(MBB, MBBI, DL, ZII->get(TargetOpcode::CFI_INSTRUCTION)) .addCFIIndex(CFIIndex); } } void SystemZELFFrameLowering::emitEpilogue(MachineFunction &MF, MachineBasicBlock &MBB) const { MachineBasicBlock::iterator MBBI = MBB.getLastNonDebugInstr(); auto *ZII = static_cast(MF.getSubtarget().getInstrInfo()); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineFrameInfo &MFFrame = MF.getFrameInfo(); // See SystemZELFFrameLowering::emitPrologue if (MF.getFunction().getCallingConv() == CallingConv::GHC) return; // Skip the return instruction. assert(MBBI->isReturn() && "Can only insert epilogue into returning blocks"); uint64_t StackSize = MFFrame.getStackSize(); if (ZFI->getRestoreGPRRegs().LowGPR) { --MBBI; unsigned Opcode = MBBI->getOpcode(); if (Opcode != SystemZ::LMG) llvm_unreachable("Expected to see callee-save register restore code"); unsigned AddrOpNo = 2; DebugLoc DL = MBBI->getDebugLoc(); uint64_t Offset = StackSize + MBBI->getOperand(AddrOpNo + 1).getImm(); unsigned NewOpcode = ZII->getOpcodeForOffset(Opcode, Offset); // If the offset is too large, use the largest stack-aligned offset // and add the rest to the base register (the stack or frame pointer). if (!NewOpcode) { uint64_t NumBytes = Offset - 0x7fff8; emitIncrement(MBB, MBBI, DL, MBBI->getOperand(AddrOpNo).getReg(), NumBytes, ZII); Offset -= NumBytes; NewOpcode = ZII->getOpcodeForOffset(Opcode, Offset); assert(NewOpcode && "No restore instruction available"); } MBBI->setDesc(ZII->get(NewOpcode)); MBBI->getOperand(AddrOpNo + 1).ChangeToImmediate(Offset); } else if (StackSize) { DebugLoc DL = MBBI->getDebugLoc(); emitIncrement(MBB, MBBI, DL, SystemZ::R15D, StackSize, ZII); } } void SystemZELFFrameLowering::inlineStackProbe( MachineFunction &MF, MachineBasicBlock &PrologMBB) const { auto *ZII = static_cast(MF.getSubtarget().getInstrInfo()); const SystemZSubtarget &STI = MF.getSubtarget(); const SystemZTargetLowering &TLI = *STI.getTargetLowering(); MachineInstr *StackAllocMI = nullptr; for (MachineInstr &MI : PrologMBB) if (MI.getOpcode() == SystemZ::PROBED_STACKALLOC) { StackAllocMI = &MI; break; } if (StackAllocMI == nullptr) return; uint64_t StackSize = StackAllocMI->getOperand(0).getImm(); const unsigned ProbeSize = TLI.getStackProbeSize(MF); uint64_t NumFullBlocks = StackSize / ProbeSize; uint64_t Residual = StackSize % ProbeSize; int64_t SPOffsetFromCFA = -SystemZMC::ELFCFAOffsetFromInitialSP; MachineBasicBlock *MBB = &PrologMBB; MachineBasicBlock::iterator MBBI = StackAllocMI; const DebugLoc DL = StackAllocMI->getDebugLoc(); // Allocate a block of Size bytes on the stack and probe it. auto allocateAndProbe = [&](MachineBasicBlock &InsMBB, MachineBasicBlock::iterator InsPt, unsigned Size, bool EmitCFI) -> void { emitIncrement(InsMBB, InsPt, DL, SystemZ::R15D, -int64_t(Size), ZII); if (EmitCFI) { SPOffsetFromCFA -= Size; buildCFAOffs(InsMBB, InsPt, DL, SPOffsetFromCFA, ZII); } // Probe by means of a volatile compare. MachineMemOperand *MMO = MF.getMachineMemOperand(MachinePointerInfo(), MachineMemOperand::MOVolatile | MachineMemOperand::MOLoad, 8, Align(1)); BuildMI(InsMBB, InsPt, DL, ZII->get(SystemZ::CG)) .addReg(SystemZ::R0D, RegState::Undef) .addReg(SystemZ::R15D).addImm(Size - 8).addReg(0) .addMemOperand(MMO); }; bool StoreBackchain = MF.getFunction().hasFnAttribute("backchain"); if (StoreBackchain) BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::LGR)) .addReg(SystemZ::R1D, RegState::Define).addReg(SystemZ::R15D); MachineBasicBlock *DoneMBB = nullptr; MachineBasicBlock *LoopMBB = nullptr; if (NumFullBlocks < 3) { // Emit unrolled probe statements. for (unsigned int i = 0; i < NumFullBlocks; i++) allocateAndProbe(*MBB, MBBI, ProbeSize, true/*EmitCFI*/); } else { // Emit a loop probing the pages. uint64_t LoopAlloc = ProbeSize * NumFullBlocks; SPOffsetFromCFA -= LoopAlloc; // Use R0D to hold the exit value. BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::LGR), SystemZ::R0D) .addReg(SystemZ::R15D); buildDefCFAReg(*MBB, MBBI, DL, SystemZ::R0D, ZII); emitIncrement(*MBB, MBBI, DL, SystemZ::R0D, -int64_t(LoopAlloc), ZII); buildCFAOffs(*MBB, MBBI, DL, -int64_t(SystemZMC::ELFCallFrameSize + LoopAlloc), ZII); DoneMBB = SystemZ::splitBlockBefore(MBBI, MBB); LoopMBB = SystemZ::emitBlockAfter(MBB); MBB->addSuccessor(LoopMBB); LoopMBB->addSuccessor(LoopMBB); LoopMBB->addSuccessor(DoneMBB); MBB = LoopMBB; allocateAndProbe(*MBB, MBB->end(), ProbeSize, false/*EmitCFI*/); BuildMI(*MBB, MBB->end(), DL, ZII->get(SystemZ::CLGR)) .addReg(SystemZ::R15D).addReg(SystemZ::R0D); BuildMI(*MBB, MBB->end(), DL, ZII->get(SystemZ::BRC)) .addImm(SystemZ::CCMASK_ICMP).addImm(SystemZ::CCMASK_CMP_GT).addMBB(MBB); MBB = DoneMBB; MBBI = DoneMBB->begin(); buildDefCFAReg(*MBB, MBBI, DL, SystemZ::R15D, ZII); } if (Residual) allocateAndProbe(*MBB, MBBI, Residual, true/*EmitCFI*/); if (StoreBackchain) BuildMI(*MBB, MBBI, DL, ZII->get(SystemZ::STG)) .addReg(SystemZ::R1D, RegState::Kill).addReg(SystemZ::R15D) .addImm(getBackchainOffset(MF)).addReg(0); StackAllocMI->eraseFromParent(); if (DoneMBB != nullptr) { // Compute the live-in lists for the new blocks. recomputeLiveIns(*DoneMBB); recomputeLiveIns(*LoopMBB); } } bool SystemZELFFrameLowering::hasFP(const MachineFunction &MF) const { return (MF.getTarget().Options.DisableFramePointerElim(MF) || MF.getFrameInfo().hasVarSizedObjects()); } StackOffset SystemZELFFrameLowering::getFrameIndexReference( const MachineFunction &MF, int FI, Register &FrameReg) const { // Our incoming SP is actually SystemZMC::ELFCallFrameSize below the CFA, so // add that difference here. StackOffset Offset = TargetFrameLowering::getFrameIndexReference(MF, FI, FrameReg); return Offset + StackOffset::getFixed(SystemZMC::ELFCallFrameSize); } unsigned SystemZELFFrameLowering::getRegSpillOffset(MachineFunction &MF, Register Reg) const { bool IsVarArg = MF.getFunction().isVarArg(); bool BackChain = MF.getFunction().hasFnAttribute("backchain"); bool SoftFloat = MF.getSubtarget().hasSoftFloat(); unsigned Offset = RegSpillOffsets[Reg]; if (usePackedStack(MF) && !(IsVarArg && !SoftFloat)) { if (SystemZ::GR64BitRegClass.contains(Reg)) // Put all GPRs at the top of the Register save area with packed // stack. Make room for the backchain if needed. Offset += BackChain ? 24 : 32; else Offset = 0; } return Offset; } int SystemZELFFrameLowering::getOrCreateFramePointerSaveIndex( MachineFunction &MF) const { SystemZMachineFunctionInfo *ZFI = MF.getInfo(); int FI = ZFI->getFramePointerSaveIndex(); if (!FI) { MachineFrameInfo &MFFrame = MF.getFrameInfo(); int Offset = getBackchainOffset(MF) - SystemZMC::ELFCallFrameSize; FI = MFFrame.CreateFixedObject(8, Offset, false); ZFI->setFramePointerSaveIndex(FI); } return FI; } bool SystemZELFFrameLowering::usePackedStack(MachineFunction &MF) const { bool HasPackedStackAttr = MF.getFunction().hasFnAttribute("packed-stack"); bool BackChain = MF.getFunction().hasFnAttribute("backchain"); bool SoftFloat = MF.getSubtarget().hasSoftFloat(); if (HasPackedStackAttr && BackChain && !SoftFloat) report_fatal_error("packed-stack + backchain + hard-float is unsupported."); bool CallConv = MF.getFunction().getCallingConv() != CallingConv::GHC; return HasPackedStackAttr && CallConv; } SystemZXPLINKFrameLowering::SystemZXPLINKFrameLowering() : SystemZFrameLowering(TargetFrameLowering::StackGrowsDown, Align(32), 0, Align(32), /* StackRealignable */ false), RegSpillOffsets(-1) { // Create a mapping from register number to save slot offset. // These offsets are relative to the start of the local are area. RegSpillOffsets.grow(SystemZ::NUM_TARGET_REGS); for (unsigned I = 0, E = array_lengthof(XPLINKSpillOffsetTable); I != E; ++I) RegSpillOffsets[XPLINKSpillOffsetTable[I].Reg] = XPLINKSpillOffsetTable[I].Offset; } // Checks if the function is a potential candidate for being a XPLeaf routine. static bool isXPLeafCandidate(const MachineFunction &MF) { const MachineFrameInfo &MFFrame = MF.getFrameInfo(); const MachineRegisterInfo &MRI = MF.getRegInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); auto *Regs = static_cast(Subtarget.getSpecialRegisters()); // If function calls other functions including alloca, then it is not a XPLeaf // routine. if (MFFrame.hasCalls()) return false; // If the function has var Sized Objects, then it is not a XPLeaf routine. if (MFFrame.hasVarSizedObjects()) return false; // If the function adjusts the stack, then it is not a XPLeaf routine. if (MFFrame.adjustsStack()) return false; // If function modifies the stack pointer register, then it is not a XPLeaf // routine. if (MRI.isPhysRegModified(Regs->getStackPointerRegister())) return false; // If function modifies the ADA register, then it is not a XPLeaf routine. if (MRI.isPhysRegModified(Regs->getAddressOfCalleeRegister())) return false; // If function modifies the return address register, then it is not a XPLeaf // routine. if (MRI.isPhysRegModified(Regs->getReturnFunctionAddressRegister())) return false; // If the backchain pointer should be stored, then it is not a XPLeaf routine. if (MF.getFunction().hasFnAttribute("backchain")) return false; // If function acquires its own stack frame, then it is not a XPLeaf routine. // At the time this function is called, only slots for local variables are // allocated, so this is a very rough estimate. if (MFFrame.estimateStackSize(MF) > 0) return false; return true; } bool SystemZXPLINKFrameLowering::assignCalleeSavedSpillSlots( MachineFunction &MF, const TargetRegisterInfo *TRI, std::vector &CSI) const { MachineFrameInfo &MFFrame = MF.getFrameInfo(); SystemZMachineFunctionInfo *MFI = MF.getInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); auto &Regs = Subtarget.getSpecialRegisters(); auto &GRRegClass = SystemZ::GR64BitRegClass; // At this point, the result of isXPLeafCandidate() is not accurate because // the size of the save area has not yet been determined. If // isXPLeafCandidate() indicates a potential leaf function, and there are no // callee-save registers, then it is indeed a leaf function, and we can early // exit. // TODO: It is possible for leaf functions to use callee-saved registers. // It can use the 0-2k range between R4 and the caller's stack frame without // acquiring its own stack frame. bool IsLeaf = CSI.empty() && isXPLeafCandidate(MF); if (IsLeaf) return true; // For non-leaf functions: // - the address of callee (entry point) register R6 must be saved CSI.push_back(CalleeSavedInfo(Regs.getAddressOfCalleeRegister())); CSI.back().setRestored(false); // The return address register R7 must be saved and restored. CSI.push_back(CalleeSavedInfo(Regs.getReturnFunctionAddressRegister())); // If the function needs a frame pointer, or if the backchain pointer should // be stored, then save the stack pointer register R4. if (hasFP(MF) || MF.getFunction().hasFnAttribute("backchain")) CSI.push_back(CalleeSavedInfo(Regs.getStackPointerRegister())); // Scan the call-saved GPRs and find the bounds of the register spill area. Register LowRestoreGPR = 0; int LowRestoreOffset = INT32_MAX; Register LowSpillGPR = 0; int LowSpillOffset = INT32_MAX; Register HighGPR = 0; int HighOffset = -1; for (auto &CS : CSI) { Register Reg = CS.getReg(); int Offset = RegSpillOffsets[Reg]; if (Offset >= 0) { if (GRRegClass.contains(Reg)) { if (LowSpillOffset > Offset) { LowSpillOffset = Offset; LowSpillGPR = Reg; } if (CS.isRestored() && LowRestoreOffset > Offset) { LowRestoreOffset = Offset; LowRestoreGPR = Reg; } if (Offset > HighOffset) { HighOffset = Offset; HighGPR = Reg; } // Non-volatile GPRs are saved in the dedicated register save area at // the bottom of the stack and are not truly part of the "normal" stack // frame. Mark the frame index as NoAlloc to indicate it as such. unsigned RegSize = 8; int FrameIdx = MFFrame.CreateFixedSpillStackObject(RegSize, Offset); CS.setFrameIdx(FrameIdx); MFFrame.setStackID(FrameIdx, TargetStackID::NoAlloc); } } else { Register Reg = CS.getReg(); const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(Reg); Align Alignment = TRI->getSpillAlign(*RC); unsigned Size = TRI->getSpillSize(*RC); Alignment = std::min(Alignment, getStackAlign()); int FrameIdx = MFFrame.CreateStackObject(Size, Alignment, true); CS.setFrameIdx(FrameIdx); } } // Save the range of call-saved registers, for use by the // prologue/epilogue inserters. if (LowRestoreGPR) MFI->setRestoreGPRRegs(LowRestoreGPR, HighGPR, LowRestoreOffset); // Save the range of call-saved registers, for use by the epilogue inserter. assert(LowSpillGPR && "Expected registers to spill"); MFI->setSpillGPRRegs(LowSpillGPR, HighGPR, LowSpillOffset); return true; } void SystemZXPLINKFrameLowering::determineCalleeSaves(MachineFunction &MF, BitVector &SavedRegs, RegScavenger *RS) const { TargetFrameLowering::determineCalleeSaves(MF, SavedRegs, RS); bool HasFP = hasFP(MF); const SystemZSubtarget &Subtarget = MF.getSubtarget(); auto &Regs = Subtarget.getSpecialRegisters(); // If the function requires a frame pointer, record that the hard // frame pointer will be clobbered. if (HasFP) SavedRegs.set(Regs.getFramePointerRegister()); } bool SystemZXPLINKFrameLowering::spillCalleeSavedRegisters( MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, ArrayRef CSI, const TargetRegisterInfo *TRI) const { if (CSI.empty()) return true; MachineFunction &MF = *MBB.getParent(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); const TargetInstrInfo *TII = Subtarget.getInstrInfo(); auto &Regs = Subtarget.getSpecialRegisters(); SystemZ::GPRRegs SpillGPRs = ZFI->getSpillGPRRegs(); DebugLoc DL; // Save GPRs if (SpillGPRs.LowGPR) { assert(SpillGPRs.LowGPR != SpillGPRs.HighGPR && "Should be saving multiple registers"); // Build an STM/STMG instruction. MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::STMG)); // Add the explicit register operands. addSavedGPR(MBB, MIB, SpillGPRs.LowGPR, false); addSavedGPR(MBB, MIB, SpillGPRs.HighGPR, false); // Add the address r4 MIB.addReg(Regs.getStackPointerRegister()); // Add the partial offset // We cannot add the actual offset as, at the stack is not finalized MIB.addImm(SpillGPRs.GPROffset); // Make sure all call-saved GPRs are included as operands and are // marked as live on entry. auto &GRRegClass = SystemZ::GR64BitRegClass; for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (GRRegClass.contains(Reg)) addSavedGPR(MBB, MIB, Reg, true); } } // Spill FPRs to the stack in the normal TargetInstrInfo way for (const CalleeSavedInfo &I : CSI) { Register Reg = I.getReg(); if (SystemZ::FP64BitRegClass.contains(Reg)) { MBB.addLiveIn(Reg); TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(), &SystemZ::FP64BitRegClass, TRI); } if (SystemZ::VR128BitRegClass.contains(Reg)) { MBB.addLiveIn(Reg); TII->storeRegToStackSlot(MBB, MBBI, Reg, true, I.getFrameIdx(), &SystemZ::VR128BitRegClass, TRI); } } return true; } bool SystemZXPLINKFrameLowering::restoreCalleeSavedRegisters( MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, MutableArrayRef CSI, const TargetRegisterInfo *TRI) const { if (CSI.empty()) return false; MachineFunction &MF = *MBB.getParent(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); const TargetInstrInfo *TII = Subtarget.getInstrInfo(); auto &Regs = Subtarget.getSpecialRegisters(); DebugLoc DL = MBBI != MBB.end() ? MBBI->getDebugLoc() : DebugLoc(); // Restore FPRs in the normal TargetInstrInfo way. for (unsigned I = 0, E = CSI.size(); I != E; ++I) { Register Reg = CSI[I].getReg(); if (SystemZ::FP64BitRegClass.contains(Reg)) TII->loadRegFromStackSlot(MBB, MBBI, Reg, CSI[I].getFrameIdx(), &SystemZ::FP64BitRegClass, TRI); if (SystemZ::VR128BitRegClass.contains(Reg)) TII->loadRegFromStackSlot(MBB, MBBI, Reg, CSI[I].getFrameIdx(), &SystemZ::VR128BitRegClass, TRI); } // Restore call-saved GPRs (but not call-clobbered varargs, which at // this point might hold return values). SystemZ::GPRRegs RestoreGPRs = ZFI->getRestoreGPRRegs(); if (RestoreGPRs.LowGPR) { assert(isInt<20>(Regs.getStackPointerBias() + RestoreGPRs.GPROffset)); if (RestoreGPRs.LowGPR == RestoreGPRs.HighGPR) // Build an LG/L instruction. BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LG), RestoreGPRs.LowGPR) .addReg(Regs.getStackPointerRegister()) .addImm(Regs.getStackPointerBias() + RestoreGPRs.GPROffset) .addReg(0); else { // Build an LMG/LM instruction. MachineInstrBuilder MIB = BuildMI(MBB, MBBI, DL, TII->get(SystemZ::LMG)); // Add the explicit register operands. MIB.addReg(RestoreGPRs.LowGPR, RegState::Define); MIB.addReg(RestoreGPRs.HighGPR, RegState::Define); // Add the address. MIB.addReg(Regs.getStackPointerRegister()); MIB.addImm(Regs.getStackPointerBias() + RestoreGPRs.GPROffset); // Do a second scan adding regs as being defined by instruction for (unsigned I = 0, E = CSI.size(); I != E; ++I) { Register Reg = CSI[I].getReg(); if (Reg > RestoreGPRs.LowGPR && Reg < RestoreGPRs.HighGPR) MIB.addReg(Reg, RegState::ImplicitDefine); } } } return true; } void SystemZXPLINKFrameLowering::emitPrologue(MachineFunction &MF, MachineBasicBlock &MBB) const { assert(&MF.front() == &MBB && "Shrink-wrapping not yet supported"); const SystemZSubtarget &Subtarget = MF.getSubtarget(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineBasicBlock::iterator MBBI = MBB.begin(); auto *ZII = static_cast(Subtarget.getInstrInfo()); auto &Regs = Subtarget.getSpecialRegisters(); MachineFrameInfo &MFFrame = MF.getFrameInfo(); MachineInstr *StoreInstr = nullptr; determineFrameLayout(MF); bool HasFP = hasFP(MF); // Debug location must be unknown since the first debug location is used // to determine the end of the prologue. DebugLoc DL; uint64_t Offset = 0; const uint64_t StackSize = MFFrame.getStackSize(); if (ZFI->getSpillGPRRegs().LowGPR) { // Skip over the GPR saves. if ((MBBI != MBB.end()) && ((MBBI->getOpcode() == SystemZ::STMG))) { const int Operand = 3; // Now we can set the offset for the operation, since now the Stack // has been finalized. Offset = Regs.getStackPointerBias() + MBBI->getOperand(Operand).getImm(); // Maximum displacement for STMG instruction. if (isInt<20>(Offset - StackSize)) Offset -= StackSize; else StoreInstr = &*MBBI; MBBI->getOperand(Operand).setImm(Offset); ++MBBI; } else llvm_unreachable("Couldn't skip over GPR saves"); } if (StackSize) { MachineBasicBlock::iterator InsertPt = StoreInstr ? StoreInstr : MBBI; // Allocate StackSize bytes. int64_t Delta = -int64_t(StackSize); // In case the STM(G) instruction also stores SP (R4), but the displacement // is too large, the SP register is manipulated first before storing, // resulting in the wrong value stored and retrieved later. In this case, we // need to temporarily save the value of SP, and store it later to memory. if (StoreInstr && HasFP) { // Insert LR r0,r4 before STMG instruction. BuildMI(MBB, InsertPt, DL, ZII->get(SystemZ::LGR)) .addReg(SystemZ::R0D, RegState::Define) .addReg(SystemZ::R4D); // Insert ST r0,xxx(,r4) after STMG instruction. BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::STG)) .addReg(SystemZ::R0D, RegState::Kill) .addReg(SystemZ::R4D) .addImm(Offset) .addReg(0); } emitIncrement(MBB, InsertPt, DL, Regs.getStackPointerRegister(), Delta, ZII); // If the requested stack size is larger than the guard page, then we need // to check if we need to call the stack extender. This requires adding a // conditional branch, but splitting the prologue block is not possible at // this point since it would invalidate the SaveBlocks / RestoreBlocks sets // of PEI in the single block function case. Build a pseudo to be handled // later by inlineStackProbe(). const uint64_t GuardPageSize = 1024 * 1024; if (StackSize > GuardPageSize) { assert(StoreInstr && "Wrong insertion point"); BuildMI(MBB, InsertPt, DL, ZII->get(SystemZ::XPLINK_STACKALLOC)); } } if (HasFP) { // Copy the base of the frame to Frame Pointer Register. BuildMI(MBB, MBBI, DL, ZII->get(SystemZ::LGR), Regs.getFramePointerRegister()) .addReg(Regs.getStackPointerRegister()); // Mark the FramePtr as live at the beginning of every block except // the entry block. (We'll have marked R8 as live on entry when // saving the GPRs.) for (MachineBasicBlock &B : llvm::drop_begin(MF)) B.addLiveIn(Regs.getFramePointerRegister()); } } void SystemZXPLINKFrameLowering::emitEpilogue(MachineFunction &MF, MachineBasicBlock &MBB) const { const SystemZSubtarget &Subtarget = MF.getSubtarget(); MachineBasicBlock::iterator MBBI = MBB.getLastNonDebugInstr(); SystemZMachineFunctionInfo *ZFI = MF.getInfo(); MachineFrameInfo &MFFrame = MF.getFrameInfo(); auto *ZII = static_cast(Subtarget.getInstrInfo()); auto &Regs = Subtarget.getSpecialRegisters(); // Skip the return instruction. assert(MBBI->isReturn() && "Can only insert epilogue into returning blocks"); uint64_t StackSize = MFFrame.getStackSize(); if (StackSize) { unsigned SPReg = Regs.getStackPointerRegister(); if (ZFI->getRestoreGPRRegs().LowGPR != SPReg) { DebugLoc DL = MBBI->getDebugLoc(); emitIncrement(MBB, MBBI, DL, SPReg, StackSize, ZII); } } } // Emit a compare of the stack pointer against the stack floor, and a call to // the LE stack extender if needed. void SystemZXPLINKFrameLowering::inlineStackProbe( MachineFunction &MF, MachineBasicBlock &PrologMBB) const { auto *ZII = static_cast(MF.getSubtarget().getInstrInfo()); MachineInstr *StackAllocMI = nullptr; for (MachineInstr &MI : PrologMBB) if (MI.getOpcode() == SystemZ::XPLINK_STACKALLOC) { StackAllocMI = &MI; break; } if (StackAllocMI == nullptr) return; MachineBasicBlock &MBB = PrologMBB; const DebugLoc DL = StackAllocMI->getDebugLoc(); // The 2nd half of block MBB after split. MachineBasicBlock *NextMBB; // Add new basic block for the call to the stack overflow function. MachineBasicBlock *StackExtMBB = MF.CreateMachineBasicBlock(MBB.getBasicBlock()); MF.push_back(StackExtMBB); // LG r3,72(,r3) BuildMI(StackExtMBB, DL, ZII->get(SystemZ::LG), SystemZ::R3D) .addReg(SystemZ::R3D) .addImm(72) .addReg(0); // BASR r3,r3 BuildMI(StackExtMBB, DL, ZII->get(SystemZ::CallBASR_STACKEXT)) .addReg(SystemZ::R3D); // LLGT r3,1208 BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::LLGT), SystemZ::R3D) .addReg(0) .addImm(1208) .addReg(0); // CG r4,64(,r3) BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::CG)) .addReg(SystemZ::R4D) .addReg(SystemZ::R3D) .addImm(64) .addReg(0); // JLL b'0100',F'37' BuildMI(MBB, StackAllocMI, DL, ZII->get(SystemZ::BRC)) .addImm(SystemZ::CCMASK_ICMP) .addImm(SystemZ::CCMASK_CMP_LT) .addMBB(StackExtMBB); NextMBB = SystemZ::splitBlockBefore(StackAllocMI, &MBB); MBB.addSuccessor(NextMBB); MBB.addSuccessor(StackExtMBB); // Add jump back from stack extension BB. BuildMI(StackExtMBB, DL, ZII->get(SystemZ::J)).addMBB(NextMBB); StackExtMBB->addSuccessor(NextMBB); StackAllocMI->eraseFromParent(); // Compute the live-in lists for the new blocks. recomputeLiveIns(*NextMBB); recomputeLiveIns(*StackExtMBB); } bool SystemZXPLINKFrameLowering::hasFP(const MachineFunction &MF) const { return (MF.getFrameInfo().hasVarSizedObjects()); } void SystemZXPLINKFrameLowering::processFunctionBeforeFrameFinalized( MachineFunction &MF, RegScavenger *RS) const { MachineFrameInfo &MFFrame = MF.getFrameInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); auto &Regs = Subtarget.getSpecialRegisters(); // Setup stack frame offset MFFrame.setOffsetAdjustment(Regs.getStackPointerBias()); } // Determines the size of the frame, and creates the deferred spill objects. void SystemZXPLINKFrameLowering::determineFrameLayout( MachineFunction &MF) const { MachineFrameInfo &MFFrame = MF.getFrameInfo(); const SystemZSubtarget &Subtarget = MF.getSubtarget(); auto *Regs = static_cast(Subtarget.getSpecialRegisters()); uint64_t StackSize = MFFrame.getStackSize(); if (StackSize == 0) return; // Add the size of the register save area and the reserved area to the size. StackSize += Regs->getCallFrameSize(); MFFrame.setStackSize(StackSize); // We now know the stack size. Create the fixed spill stack objects for the // register save area now. This has no impact on the stack frame layout, as // this is already computed. However, it makes sure that all callee saved // registers have a valid frame index assigned. const unsigned RegSize = MF.getDataLayout().getPointerSize(); for (auto &CS : MFFrame.getCalleeSavedInfo()) { int Offset = RegSpillOffsets[CS.getReg()]; if (Offset >= 0) CS.setFrameIdx( MFFrame.CreateFixedSpillStackObject(RegSize, Offset - StackSize)); } }