1 //===- StackMaps.cpp ------------------------------------------------------===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 #include "llvm/CodeGen/StackMaps.h"
11 #include "llvm/ADT/DenseMapInfo.h"
12 #include "llvm/ADT/STLExtras.h"
13 #include "llvm/ADT/Twine.h"
14 #include "llvm/CodeGen/AsmPrinter.h"
15 #include "llvm/CodeGen/MachineFrameInfo.h"
16 #include "llvm/CodeGen/MachineFunction.h"
17 #include "llvm/CodeGen/MachineInstr.h"
18 #include "llvm/CodeGen/MachineOperand.h"
19 #include "llvm/CodeGen/TargetOpcodes.h"
20 #include "llvm/CodeGen/TargetRegisterInfo.h"
21 #include "llvm/CodeGen/TargetSubtargetInfo.h"
22 #include "llvm/IR/DataLayout.h"
23 #include "llvm/MC/MCContext.h"
24 #include "llvm/MC/MCExpr.h"
25 #include "llvm/MC/MCObjectFileInfo.h"
26 #include "llvm/MC/MCRegisterInfo.h"
27 #include "llvm/MC/MCStreamer.h"
28 #include "llvm/Support/CommandLine.h"
29 #include "llvm/Support/Debug.h"
30 #include "llvm/Support/ErrorHandling.h"
31 #include "llvm/Support/MathExtras.h"
32 #include "llvm/Support/raw_ostream.h"
41 #define DEBUG_TYPE "stackmaps"
43 static cl::opt<int> StackMapVersion(
44 "stackmap-version", cl::init(3), cl::Hidden,
45 cl::desc("Specify the stackmap encoding version (default = 3)"));
47 const char *StackMaps::WSMP = "Stack Maps: ";
49 StackMapOpers::StackMapOpers(const MachineInstr *MI)
51 assert(getVarIdx() <= MI->getNumOperands() &&
52 "invalid stackmap definition");
55 PatchPointOpers::PatchPointOpers(const MachineInstr *MI)
56 : MI(MI), HasDef(MI->getOperand(0).isReg() && MI->getOperand(0).isDef() &&
57 !MI->getOperand(0).isImplicit()) {
59 unsigned CheckStartIdx = 0, e = MI->getNumOperands();
60 while (CheckStartIdx < e && MI->getOperand(CheckStartIdx).isReg() &&
61 MI->getOperand(CheckStartIdx).isDef() &&
62 !MI->getOperand(CheckStartIdx).isImplicit())
65 assert(getMetaIdx() == CheckStartIdx &&
66 "Unexpected additional definition in Patchpoint intrinsic.");
70 unsigned PatchPointOpers::getNextScratchIdx(unsigned StartIdx) const {
72 StartIdx = getVarIdx();
74 // Find the next scratch register (implicit def and early clobber)
75 unsigned ScratchIdx = StartIdx, e = MI->getNumOperands();
76 while (ScratchIdx < e &&
77 !(MI->getOperand(ScratchIdx).isReg() &&
78 MI->getOperand(ScratchIdx).isDef() &&
79 MI->getOperand(ScratchIdx).isImplicit() &&
80 MI->getOperand(ScratchIdx).isEarlyClobber()))
83 assert(ScratchIdx != e && "No scratch register available");
87 StackMaps::StackMaps(AsmPrinter &AP) : AP(AP) {
88 if (StackMapVersion != 3)
89 llvm_unreachable("Unsupported stackmap version!");
92 /// Go up the super-register chain until we hit a valid dwarf register number.
93 static unsigned getDwarfRegNum(unsigned Reg, const TargetRegisterInfo *TRI) {
94 int RegNum = TRI->getDwarfRegNum(Reg, false);
95 for (MCSuperRegIterator SR(Reg, TRI); SR.isValid() && RegNum < 0; ++SR)
96 RegNum = TRI->getDwarfRegNum(*SR, false);
98 assert(RegNum >= 0 && "Invalid Dwarf register number.");
99 return (unsigned)RegNum;
102 MachineInstr::const_mop_iterator
103 StackMaps::parseOperand(MachineInstr::const_mop_iterator MOI,
104 MachineInstr::const_mop_iterator MOE, LocationVec &Locs,
105 LiveOutVec &LiveOuts) const {
106 const TargetRegisterInfo *TRI = AP.MF->getSubtarget().getRegisterInfo();
108 switch (MOI->getImm()) {
110 llvm_unreachable("Unrecognized operand type.");
111 case StackMaps::DirectMemRefOp: {
112 auto &DL = AP.MF->getDataLayout();
114 unsigned Size = DL.getPointerSizeInBits();
115 assert((Size % 8) == 0 && "Need pointer size in bytes.");
117 unsigned Reg = (++MOI)->getReg();
118 int64_t Imm = (++MOI)->getImm();
119 Locs.emplace_back(StackMaps::Location::Direct, Size,
120 getDwarfRegNum(Reg, TRI), Imm);
123 case StackMaps::IndirectMemRefOp: {
124 int64_t Size = (++MOI)->getImm();
125 assert(Size > 0 && "Need a valid size for indirect memory locations.");
126 unsigned Reg = (++MOI)->getReg();
127 int64_t Imm = (++MOI)->getImm();
128 Locs.emplace_back(StackMaps::Location::Indirect, Size,
129 getDwarfRegNum(Reg, TRI), Imm);
132 case StackMaps::ConstantOp: {
134 assert(MOI->isImm() && "Expected constant operand.");
135 int64_t Imm = MOI->getImm();
136 Locs.emplace_back(Location::Constant, sizeof(int64_t), 0, Imm);
143 // The physical register number will ultimately be encoded as a DWARF regno.
144 // The stack map also records the size of a spill slot that can hold the
145 // register content. (The runtime can track the actual size of the data type
148 // Skip implicit registers (this includes our scratch registers)
149 if (MOI->isImplicit())
152 assert(TargetRegisterInfo::isPhysicalRegister(MOI->getReg()) &&
153 "Virtreg operands should have been rewritten before now.");
154 const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(MOI->getReg());
155 assert(!MOI->getSubReg() && "Physical subreg still around.");
158 unsigned DwarfRegNum = getDwarfRegNum(MOI->getReg(), TRI);
159 unsigned LLVMRegNum = TRI->getLLVMRegNum(DwarfRegNum, false);
160 unsigned SubRegIdx = TRI->getSubRegIndex(LLVMRegNum, MOI->getReg());
162 Offset = TRI->getSubRegIdxOffset(SubRegIdx);
164 Locs.emplace_back(Location::Register, TRI->getSpillSize(*RC),
165 DwarfRegNum, Offset);
169 if (MOI->isRegLiveOut())
170 LiveOuts = parseRegisterLiveOutMask(MOI->getRegLiveOut());
175 void StackMaps::print(raw_ostream &OS) {
176 const TargetRegisterInfo *TRI =
177 AP.MF ? AP.MF->getSubtarget().getRegisterInfo() : nullptr;
178 OS << WSMP << "callsites:\n";
179 for (const auto &CSI : CSInfos) {
180 const LocationVec &CSLocs = CSI.Locations;
181 const LiveOutVec &LiveOuts = CSI.LiveOuts;
183 OS << WSMP << "callsite " << CSI.ID << "\n";
184 OS << WSMP << " has " << CSLocs.size() << " locations\n";
187 for (const auto &Loc : CSLocs) {
188 OS << WSMP << "\t\tLoc " << Idx << ": ";
190 case Location::Unprocessed:
191 OS << "<Unprocessed operand>";
193 case Location::Register:
196 OS << printReg(Loc.Reg, TRI);
200 case Location::Direct:
203 OS << printReg(Loc.Reg, TRI);
207 OS << " + " << Loc.Offset;
209 case Location::Indirect:
212 OS << printReg(Loc.Reg, TRI);
215 OS << "+" << Loc.Offset;
217 case Location::Constant:
218 OS << "Constant " << Loc.Offset;
220 case Location::ConstantIndex:
221 OS << "Constant Index " << Loc.Offset;
224 OS << "\t[encoding: .byte " << Loc.Type << ", .byte 0"
225 << ", .short " << Loc.Size << ", .short " << Loc.Reg << ", .short 0"
226 << ", .int " << Loc.Offset << "]\n";
230 OS << WSMP << "\thas " << LiveOuts.size() << " live-out registers\n";
233 for (const auto &LO : LiveOuts) {
234 OS << WSMP << "\t\tLO " << Idx << ": ";
236 OS << printReg(LO.Reg, TRI);
239 OS << "\t[encoding: .short " << LO.DwarfRegNum << ", .byte 0, .byte "
246 /// Create a live-out register record for the given register Reg.
247 StackMaps::LiveOutReg
248 StackMaps::createLiveOutReg(unsigned Reg, const TargetRegisterInfo *TRI) const {
249 unsigned DwarfRegNum = getDwarfRegNum(Reg, TRI);
250 unsigned Size = TRI->getSpillSize(*TRI->getMinimalPhysRegClass(Reg));
251 return LiveOutReg(Reg, DwarfRegNum, Size);
254 /// Parse the register live-out mask and return a vector of live-out registers
255 /// that need to be recorded in the stackmap.
256 StackMaps::LiveOutVec
257 StackMaps::parseRegisterLiveOutMask(const uint32_t *Mask) const {
258 assert(Mask && "No register mask specified");
259 const TargetRegisterInfo *TRI = AP.MF->getSubtarget().getRegisterInfo();
262 // Create a LiveOutReg for each bit that is set in the register mask.
263 for (unsigned Reg = 0, NumRegs = TRI->getNumRegs(); Reg != NumRegs; ++Reg)
264 if ((Mask[Reg / 32] >> Reg % 32) & 1)
265 LiveOuts.push_back(createLiveOutReg(Reg, TRI));
267 // We don't need to keep track of a register if its super-register is already
268 // in the list. Merge entries that refer to the same dwarf register and use
269 // the maximum size that needs to be spilled.
271 llvm::sort(LiveOuts.begin(), LiveOuts.end(),
272 [](const LiveOutReg &LHS, const LiveOutReg &RHS) {
273 // Only sort by the dwarf register number.
274 return LHS.DwarfRegNum < RHS.DwarfRegNum;
277 for (auto I = LiveOuts.begin(), E = LiveOuts.end(); I != E; ++I) {
278 for (auto II = std::next(I); II != E; ++II) {
279 if (I->DwarfRegNum != II->DwarfRegNum) {
280 // Skip all the now invalid entries.
284 I->Size = std::max(I->Size, II->Size);
285 if (TRI->isSuperRegister(I->Reg, II->Reg))
287 II->Reg = 0; // mark for deletion.
292 llvm::remove_if(LiveOuts,
293 [](const LiveOutReg &LO) { return LO.Reg == 0; }),
299 void StackMaps::recordStackMapOpers(const MachineInstr &MI, uint64_t ID,
300 MachineInstr::const_mop_iterator MOI,
301 MachineInstr::const_mop_iterator MOE,
303 MCContext &OutContext = AP.OutStreamer->getContext();
304 MCSymbol *MILabel = OutContext.createTempSymbol();
305 AP.OutStreamer->EmitLabel(MILabel);
307 LocationVec Locations;
311 assert(PatchPointOpers(&MI).hasDef() && "Stackmap has no return value.");
312 parseOperand(MI.operands_begin(), std::next(MI.operands_begin()), Locations,
318 MOI = parseOperand(MOI, MOE, Locations, LiveOuts);
321 // Move large constants into the constant pool.
322 for (auto &Loc : Locations) {
323 // Constants are encoded as sign-extended integers.
324 // -1 is directly encoded as .long 0xFFFFFFFF with no constant pool.
325 if (Loc.Type == Location::Constant && !isInt<32>(Loc.Offset)) {
326 Loc.Type = Location::ConstantIndex;
327 // ConstPool is intentionally a MapVector of 'uint64_t's (as
328 // opposed to 'int64_t's). We should never be in a situation
329 // where we have to insert either the tombstone or the empty
330 // keys into a map, and for a DenseMap<uint64_t, T> these are
331 // (uint64_t)0 and (uint64_t)-1. They can be and are
332 // represented using 32 bit integers.
333 assert((uint64_t)Loc.Offset != DenseMapInfo<uint64_t>::getEmptyKey() &&
334 (uint64_t)Loc.Offset !=
335 DenseMapInfo<uint64_t>::getTombstoneKey() &&
336 "empty and tombstone keys should fit in 32 bits!");
337 auto Result = ConstPool.insert(std::make_pair(Loc.Offset, Loc.Offset));
338 Loc.Offset = Result.first - ConstPool.begin();
342 // Create an expression to calculate the offset of the callsite from function
344 const MCExpr *CSOffsetExpr = MCBinaryExpr::createSub(
345 MCSymbolRefExpr::create(MILabel, OutContext),
346 MCSymbolRefExpr::create(AP.CurrentFnSymForSize, OutContext), OutContext);
348 CSInfos.emplace_back(CSOffsetExpr, ID, std::move(Locations),
349 std::move(LiveOuts));
351 // Record the stack size of the current function and update callsite count.
352 const MachineFrameInfo &MFI = AP.MF->getFrameInfo();
353 const TargetRegisterInfo *RegInfo = AP.MF->getSubtarget().getRegisterInfo();
354 bool HasDynamicFrameSize =
355 MFI.hasVarSizedObjects() || RegInfo->needsStackRealignment(*(AP.MF));
356 uint64_t FrameSize = HasDynamicFrameSize ? UINT64_MAX : MFI.getStackSize();
358 auto CurrentIt = FnInfos.find(AP.CurrentFnSym);
359 if (CurrentIt != FnInfos.end())
360 CurrentIt->second.RecordCount++;
362 FnInfos.insert(std::make_pair(AP.CurrentFnSym, FunctionInfo(FrameSize)));
365 void StackMaps::recordStackMap(const MachineInstr &MI) {
366 assert(MI.getOpcode() == TargetOpcode::STACKMAP && "expected stackmap");
368 StackMapOpers opers(&MI);
369 const int64_t ID = MI.getOperand(PatchPointOpers::IDPos).getImm();
370 recordStackMapOpers(MI, ID, std::next(MI.operands_begin(), opers.getVarIdx()),
374 void StackMaps::recordPatchPoint(const MachineInstr &MI) {
375 assert(MI.getOpcode() == TargetOpcode::PATCHPOINT && "expected patchpoint");
377 PatchPointOpers opers(&MI);
378 const int64_t ID = opers.getID();
379 auto MOI = std::next(MI.operands_begin(), opers.getStackMapStartIdx());
380 recordStackMapOpers(MI, ID, MOI, MI.operands_end(),
381 opers.isAnyReg() && opers.hasDef());
385 auto &Locations = CSInfos.back().Locations;
386 if (opers.isAnyReg()) {
387 unsigned NArgs = opers.getNumCallArgs();
388 for (unsigned i = 0, e = (opers.hasDef() ? NArgs + 1 : NArgs); i != e; ++i)
389 assert(Locations[i].Type == Location::Register &&
390 "anyreg arg must be in reg.");
395 void StackMaps::recordStatepoint(const MachineInstr &MI) {
396 assert(MI.getOpcode() == TargetOpcode::STATEPOINT && "expected statepoint");
398 StatepointOpers opers(&MI);
399 // Record all the deopt and gc operands (they're contiguous and run from the
400 // initial index to the end of the operand list)
401 const unsigned StartIdx = opers.getVarIdx();
402 recordStackMapOpers(MI, opers.getID(), MI.operands_begin() + StartIdx,
403 MI.operands_end(), false);
406 /// Emit the stackmap header.
409 /// uint8 : Stack Map Version (currently 2)
410 /// uint8 : Reserved (expected to be 0)
411 /// uint16 : Reserved (expected to be 0)
413 /// uint32 : NumFunctions
414 /// uint32 : NumConstants
415 /// uint32 : NumRecords
416 void StackMaps::emitStackmapHeader(MCStreamer &OS) {
418 OS.EmitIntValue(StackMapVersion, 1); // Version.
419 OS.EmitIntValue(0, 1); // Reserved.
420 OS.EmitIntValue(0, 2); // Reserved.
423 LLVM_DEBUG(dbgs() << WSMP << "#functions = " << FnInfos.size() << '\n');
424 OS.EmitIntValue(FnInfos.size(), 4);
426 LLVM_DEBUG(dbgs() << WSMP << "#constants = " << ConstPool.size() << '\n');
427 OS.EmitIntValue(ConstPool.size(), 4);
429 LLVM_DEBUG(dbgs() << WSMP << "#callsites = " << CSInfos.size() << '\n');
430 OS.EmitIntValue(CSInfos.size(), 4);
433 /// Emit the function frame record for each function.
435 /// StkSizeRecord[NumFunctions] {
436 /// uint64 : Function Address
437 /// uint64 : Stack Size
438 /// uint64 : Record Count
440 void StackMaps::emitFunctionFrameRecords(MCStreamer &OS) {
441 // Function Frame records.
442 LLVM_DEBUG(dbgs() << WSMP << "functions:\n");
443 for (auto const &FR : FnInfos) {
444 LLVM_DEBUG(dbgs() << WSMP << "function addr: " << FR.first
445 << " frame size: " << FR.second.StackSize
446 << " callsite count: " << FR.second.RecordCount << '\n');
447 OS.EmitSymbolValue(FR.first, 8);
448 OS.EmitIntValue(FR.second.StackSize, 8);
449 OS.EmitIntValue(FR.second.RecordCount, 8);
453 /// Emit the constant pool.
455 /// int64 : Constants[NumConstants]
456 void StackMaps::emitConstantPoolEntries(MCStreamer &OS) {
457 // Constant pool entries.
458 LLVM_DEBUG(dbgs() << WSMP << "constants:\n");
459 for (const auto &ConstEntry : ConstPool) {
460 LLVM_DEBUG(dbgs() << WSMP << ConstEntry.second << '\n');
461 OS.EmitIntValue(ConstEntry.second, 8);
465 /// Emit the callsite info for each callsite.
467 /// StkMapRecord[NumRecords] {
468 /// uint64 : PatchPoint ID
469 /// uint32 : Instruction Offset
470 /// uint16 : Reserved (record flags)
471 /// uint16 : NumLocations
472 /// Location[NumLocations] {
473 /// uint8 : Register | Direct | Indirect | Constant | ConstantIndex
474 /// uint8 : Size in Bytes
475 /// uint16 : Dwarf RegNum
479 /// uint16 : NumLiveOuts
480 /// LiveOuts[NumLiveOuts] {
481 /// uint16 : Dwarf RegNum
483 /// uint8 : Size in Bytes
485 /// uint32 : Padding (only if required to align to 8 byte)
488 /// Location Encoding, Type, Value:
489 /// 0x1, Register, Reg (value in register)
490 /// 0x2, Direct, Reg + Offset (frame index)
491 /// 0x3, Indirect, [Reg + Offset] (spilled value)
492 /// 0x4, Constant, Offset (small constant)
493 /// 0x5, ConstIndex, Constants[Offset] (large constant)
494 void StackMaps::emitCallsiteEntries(MCStreamer &OS) {
495 LLVM_DEBUG(print(dbgs()));
497 for (const auto &CSI : CSInfos) {
498 const LocationVec &CSLocs = CSI.Locations;
499 const LiveOutVec &LiveOuts = CSI.LiveOuts;
501 // Verify stack map entry. It's better to communicate a problem to the
502 // runtime than crash in case of in-process compilation. Currently, we do
503 // simple overflow checks, but we may eventually communicate other
504 // compilation errors this way.
505 if (CSLocs.size() > UINT16_MAX || LiveOuts.size() > UINT16_MAX) {
506 OS.EmitIntValue(UINT64_MAX, 8); // Invalid ID.
507 OS.EmitValue(CSI.CSOffsetExpr, 4);
508 OS.EmitIntValue(0, 2); // Reserved.
509 OS.EmitIntValue(0, 2); // 0 locations.
510 OS.EmitIntValue(0, 2); // padding.
511 OS.EmitIntValue(0, 2); // 0 live-out registers.
512 OS.EmitIntValue(0, 4); // padding.
516 OS.EmitIntValue(CSI.ID, 8);
517 OS.EmitValue(CSI.CSOffsetExpr, 4);
519 // Reserved for flags.
520 OS.EmitIntValue(0, 2);
521 OS.EmitIntValue(CSLocs.size(), 2);
523 for (const auto &Loc : CSLocs) {
524 OS.EmitIntValue(Loc.Type, 1);
525 OS.EmitIntValue(0, 1); // Reserved
526 OS.EmitIntValue(Loc.Size, 2);
527 OS.EmitIntValue(Loc.Reg, 2);
528 OS.EmitIntValue(0, 2); // Reserved
529 OS.EmitIntValue(Loc.Offset, 4);
532 // Emit alignment to 8 byte.
533 OS.EmitValueToAlignment(8);
535 // Num live-out registers and padding to align to 4 byte.
536 OS.EmitIntValue(0, 2);
537 OS.EmitIntValue(LiveOuts.size(), 2);
539 for (const auto &LO : LiveOuts) {
540 OS.EmitIntValue(LO.DwarfRegNum, 2);
541 OS.EmitIntValue(0, 1);
542 OS.EmitIntValue(LO.Size, 1);
544 // Emit alignment to 8 byte.
545 OS.EmitValueToAlignment(8);
549 /// Serialize the stackmap data.
550 void StackMaps::serializeToStackMapSection() {
552 // Bail out if there's no stack map data.
553 assert((!CSInfos.empty() || ConstPool.empty()) &&
554 "Expected empty constant pool too!");
555 assert((!CSInfos.empty() || FnInfos.empty()) &&
556 "Expected empty function record too!");
560 MCContext &OutContext = AP.OutStreamer->getContext();
561 MCStreamer &OS = *AP.OutStreamer;
563 // Create the section.
564 MCSection *StackMapSection =
565 OutContext.getObjectFileInfo()->getStackMapSection();
566 OS.SwitchSection(StackMapSection);
568 // Emit a dummy symbol to force section inclusion.
569 OS.EmitLabel(OutContext.getOrCreateSymbol(Twine("__LLVM_StackMaps")));
572 LLVM_DEBUG(dbgs() << "********** Stack Map Output **********\n");
573 emitStackmapHeader(OS);
574 emitFunctionFrameRecords(OS);
575 emitConstantPoolEntries(OS);
576 emitCallsiteEntries(OS);