1 //===-- CodeGenFunction.h - Per-Function state for LLVM CodeGen -*- C++ -*-===//
3 // The LLVM Compiler Infrastructure
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
8 //===----------------------------------------------------------------------===//
10 // This is the internal per-function state used for llvm translation.
12 //===----------------------------------------------------------------------===//
14 #ifndef CLANG_CODEGEN_CODEGENFUNCTION_H
15 #define CLANG_CODEGEN_CODEGENFUNCTION_H
17 #include "clang/AST/Type.h"
18 #include "clang/AST/ExprCXX.h"
19 #include "clang/AST/ExprObjC.h"
20 #include "clang/AST/CharUnits.h"
21 #include "clang/Basic/ABI.h"
22 #include "clang/Basic/TargetInfo.h"
23 #include "llvm/ADT/DenseMap.h"
24 #include "llvm/ADT/SmallVector.h"
25 #include "llvm/Support/ValueHandle.h"
26 #include "CodeGenModule.h"
27 #include "CGBuilder.h"
44 class CXXDestructorDecl;
45 class CXXForRangeStmt;
49 class EnumConstantDecl;
51 class FunctionProtoType;
53 class ObjCContainerDecl;
54 class ObjCInterfaceDecl;
57 class ObjCImplementationDecl;
58 class ObjCPropertyImplDecl;
60 class TargetCodeGenInfo;
62 class ObjCForCollectionStmt;
64 class ObjCAtThrowStmt;
65 class ObjCAtSynchronizedStmt;
75 class BlockFieldFlags;
77 /// A branch fixup. These are required when emitting a goto to a
78 /// label which hasn't been emitted yet. The goto is optimistically
79 /// emitted as a branch to the basic block for the label, and (if it
80 /// occurs in a scope with non-trivial cleanups) a fixup is added to
81 /// the innermost cleanup. When a (normal) cleanup is popped, any
82 /// unresolved fixups in that scope are threaded through the cleanup.
84 /// The block containing the terminator which needs to be modified
85 /// into a switch if this fixup is resolved into the current scope.
86 /// If null, LatestBranch points directly to the destination.
87 llvm::BasicBlock *OptimisticBranchBlock;
89 /// The ultimate destination of the branch.
91 /// This can be set to null to indicate that this fixup was
92 /// successfully resolved.
93 llvm::BasicBlock *Destination;
95 /// The destination index value.
96 unsigned DestinationIndex;
98 /// The initial branch of the fixup.
99 llvm::BranchInst *InitialBranch;
102 template <class T> struct InvariantValue {
104 typedef T saved_type;
105 static bool needsSaving(type value) { return false; }
106 static saved_type save(CodeGenFunction &CGF, type value) { return value; }
107 static type restore(CodeGenFunction &CGF, saved_type value) { return value; }
110 /// A metaprogramming class for ensuring that a value will dominate an
111 /// arbitrary position in a function.
112 template <class T> struct DominatingValue : InvariantValue<T> {};
114 template <class T, bool mightBeInstruction =
115 llvm::is_base_of<llvm::Value, T>::value &&
116 !llvm::is_base_of<llvm::Constant, T>::value &&
117 !llvm::is_base_of<llvm::BasicBlock, T>::value>
118 struct DominatingPointer;
119 template <class T> struct DominatingPointer<T,false> : InvariantValue<T*> {};
120 // template <class T> struct DominatingPointer<T,true> at end of file
122 template <class T> struct DominatingValue<T*> : DominatingPointer<T> {};
127 NormalAndEHCleanup = EHCleanup | NormalCleanup,
129 InactiveCleanup = 0x4,
130 InactiveEHCleanup = EHCleanup | InactiveCleanup,
131 InactiveNormalCleanup = NormalCleanup | InactiveCleanup,
132 InactiveNormalAndEHCleanup = NormalAndEHCleanup | InactiveCleanup
135 /// A stack of scopes which respond to exceptions, including cleanups
136 /// and catch blocks.
139 /// A saved depth on the scope stack. This is necessary because
140 /// pushing scopes onto the stack invalidates iterators.
141 class stable_iterator {
142 friend class EHScopeStack;
144 /// Offset from StartOfData to EndOfBuffer.
147 stable_iterator(ptrdiff_t Size) : Size(Size) {}
150 static stable_iterator invalid() { return stable_iterator(-1); }
151 stable_iterator() : Size(-1) {}
153 bool isValid() const { return Size >= 0; }
155 /// Returns true if this scope encloses I.
156 /// Returns false if I is invalid.
157 /// This scope must be valid.
158 bool encloses(stable_iterator I) const { return Size <= I.Size; }
160 /// Returns true if this scope strictly encloses I: that is,
161 /// if it encloses I and is not I.
162 /// Returns false is I is invalid.
163 /// This scope must be valid.
164 bool strictlyEncloses(stable_iterator I) const { return Size < I.Size; }
166 friend bool operator==(stable_iterator A, stable_iterator B) {
167 return A.Size == B.Size;
169 friend bool operator!=(stable_iterator A, stable_iterator B) {
170 return A.Size != B.Size;
174 /// Information for lazily generating a cleanup. Subclasses must be
175 /// POD-like: cleanups will not be destructed, and they will be
176 /// allocated on the cleanup stack and freely copied and moved
179 /// Cleanup implementations should generally be declared in an
180 /// anonymous namespace.
183 // Anchor the construction vtable. We use the destructor because
184 // gcc gives an obnoxious warning if there are virtual methods
185 // with an accessible non-virtual destructor. Unfortunately,
186 // declaring this destructor makes it non-trivial, but there
187 // doesn't seem to be any other way around this warning.
189 // This destructor will never be called.
192 /// Emit the cleanup. For normal cleanups, this is run in the
193 /// same EH context as when the cleanup was pushed, i.e. the
194 /// immediately-enclosing context of the cleanup scope. For
195 /// EH cleanups, this is run in a terminate context.
197 // \param IsForEHCleanup true if this is for an EH cleanup, false
198 /// if for a normal cleanup.
199 virtual void Emit(CodeGenFunction &CGF, bool IsForEHCleanup) = 0;
202 /// UnconditionalCleanupN stores its N parameters and just passes
203 /// them to the real cleanup function.
204 template <class T, class A0>
205 class UnconditionalCleanup1 : public Cleanup {
208 UnconditionalCleanup1(A0 a0) : a0(a0) {}
209 void Emit(CodeGenFunction &CGF, bool IsForEHCleanup) {
210 T::Emit(CGF, IsForEHCleanup, a0);
214 template <class T, class A0, class A1>
215 class UnconditionalCleanup2 : public Cleanup {
218 UnconditionalCleanup2(A0 a0, A1 a1) : a0(a0), a1(a1) {}
219 void Emit(CodeGenFunction &CGF, bool IsForEHCleanup) {
220 T::Emit(CGF, IsForEHCleanup, a0, a1);
224 /// ConditionalCleanupN stores the saved form of its N parameters,
225 /// then restores them and performs the cleanup.
226 template <class T, class A0>
227 class ConditionalCleanup1 : public Cleanup {
228 typedef typename DominatingValue<A0>::saved_type A0_saved;
231 void Emit(CodeGenFunction &CGF, bool IsForEHCleanup) {
232 A0 a0 = DominatingValue<A0>::restore(CGF, a0_saved);
233 T::Emit(CGF, IsForEHCleanup, a0);
237 ConditionalCleanup1(A0_saved a0)
241 template <class T, class A0, class A1>
242 class ConditionalCleanup2 : public Cleanup {
243 typedef typename DominatingValue<A0>::saved_type A0_saved;
244 typedef typename DominatingValue<A1>::saved_type A1_saved;
248 void Emit(CodeGenFunction &CGF, bool IsForEHCleanup) {
249 A0 a0 = DominatingValue<A0>::restore(CGF, a0_saved);
250 A1 a1 = DominatingValue<A1>::restore(CGF, a1_saved);
251 T::Emit(CGF, IsForEHCleanup, a0, a1);
255 ConditionalCleanup2(A0_saved a0, A1_saved a1)
256 : a0_saved(a0), a1_saved(a1) {}
260 // The implementation for this class is in CGException.h and
261 // CGException.cpp; the definition is here because it's used as a
262 // member of CodeGenFunction.
264 /// The start of the scope-stack buffer, i.e. the allocated pointer
265 /// for the buffer. All of these pointers are either simultaneously
266 /// null or simultaneously valid.
269 /// The end of the buffer.
272 /// The first valid entry in the buffer.
275 /// The innermost normal cleanup on the stack.
276 stable_iterator InnermostNormalCleanup;
278 /// The innermost EH cleanup on the stack.
279 stable_iterator InnermostEHCleanup;
281 /// The number of catches on the stack.
284 /// The current EH destination index. Reset to FirstCatchIndex
285 /// whenever the last EH cleanup is popped.
286 unsigned NextEHDestIndex;
287 enum { FirstEHDestIndex = 1 };
289 /// The current set of branch fixups. A branch fixup is a jump to
290 /// an as-yet unemitted label, i.e. a label for which we don't yet
291 /// know the EH stack depth. Whenever we pop a cleanup, we have
292 /// to thread all the current branch fixups through it.
294 /// Fixups are recorded as the Use of the respective branch or
295 /// switch statement. The use points to the final destination.
296 /// When popping out of a cleanup, these uses are threaded through
297 /// the cleanup and adjusted to point to the new cleanup.
299 /// Note that branches are allowed to jump into protected scopes
300 /// in certain situations; e.g. the following code is legal:
301 /// struct A { ~A(); }; // trivial ctor, non-trivial dtor
306 llvm::SmallVector<BranchFixup, 8> BranchFixups;
308 char *allocate(size_t Size);
310 void *pushCleanup(CleanupKind K, size_t DataSize);
313 EHScopeStack() : StartOfBuffer(0), EndOfBuffer(0), StartOfData(0),
314 InnermostNormalCleanup(stable_end()),
315 InnermostEHCleanup(stable_end()),
316 CatchDepth(0), NextEHDestIndex(FirstEHDestIndex) {}
317 ~EHScopeStack() { delete[] StartOfBuffer; }
319 // Variadic templates would make this not terrible.
321 /// Push a lazily-created cleanup on the stack.
323 void pushCleanup(CleanupKind Kind) {
324 void *Buffer = pushCleanup(Kind, sizeof(T));
325 Cleanup *Obj = new(Buffer) T();
329 /// Push a lazily-created cleanup on the stack.
330 template <class T, class A0>
331 void pushCleanup(CleanupKind Kind, A0 a0) {
332 void *Buffer = pushCleanup(Kind, sizeof(T));
333 Cleanup *Obj = new(Buffer) T(a0);
337 /// Push a lazily-created cleanup on the stack.
338 template <class T, class A0, class A1>
339 void pushCleanup(CleanupKind Kind, A0 a0, A1 a1) {
340 void *Buffer = pushCleanup(Kind, sizeof(T));
341 Cleanup *Obj = new(Buffer) T(a0, a1);
345 /// Push a lazily-created cleanup on the stack.
346 template <class T, class A0, class A1, class A2>
347 void pushCleanup(CleanupKind Kind, A0 a0, A1 a1, A2 a2) {
348 void *Buffer = pushCleanup(Kind, sizeof(T));
349 Cleanup *Obj = new(Buffer) T(a0, a1, a2);
353 /// Push a lazily-created cleanup on the stack.
354 template <class T, class A0, class A1, class A2, class A3>
355 void pushCleanup(CleanupKind Kind, A0 a0, A1 a1, A2 a2, A3 a3) {
356 void *Buffer = pushCleanup(Kind, sizeof(T));
357 Cleanup *Obj = new(Buffer) T(a0, a1, a2, a3);
361 /// Push a lazily-created cleanup on the stack.
362 template <class T, class A0, class A1, class A2, class A3, class A4>
363 void pushCleanup(CleanupKind Kind, A0 a0, A1 a1, A2 a2, A3 a3, A4 a4) {
364 void *Buffer = pushCleanup(Kind, sizeof(T));
365 Cleanup *Obj = new(Buffer) T(a0, a1, a2, a3, a4);
369 // Feel free to add more variants of the following:
371 /// Push a cleanup with non-constant storage requirements on the
372 /// stack. The cleanup type must provide an additional static method:
373 /// static size_t getExtraSize(size_t);
374 /// The argument to this method will be the value N, which will also
375 /// be passed as the first argument to the constructor.
377 /// The data stored in the extra storage must obey the same
378 /// restrictions as normal cleanup member data.
380 /// The pointer returned from this method is valid until the cleanup
381 /// stack is modified.
382 template <class T, class A0, class A1, class A2>
383 T *pushCleanupWithExtra(CleanupKind Kind, size_t N, A0 a0, A1 a1, A2 a2) {
384 void *Buffer = pushCleanup(Kind, sizeof(T) + T::getExtraSize(N));
385 return new (Buffer) T(N, a0, a1, a2);
388 /// Pops a cleanup scope off the stack. This should only be called
389 /// by CodeGenFunction::PopCleanupBlock.
392 /// Push a set of catch handlers on the stack. The catch is
393 /// uninitialized and will need to have the given number of handlers
395 class EHCatchScope *pushCatch(unsigned NumHandlers);
397 /// Pops a catch scope off the stack.
400 /// Push an exceptions filter on the stack.
401 class EHFilterScope *pushFilter(unsigned NumFilters);
403 /// Pops an exceptions filter off the stack.
406 /// Push a terminate handler on the stack.
407 void pushTerminate();
409 /// Pops a terminate handler off the stack.
412 /// Determines whether the exception-scopes stack is empty.
413 bool empty() const { return StartOfData == EndOfBuffer; }
415 bool requiresLandingPad() const {
416 return (CatchDepth || hasEHCleanups());
419 /// Determines whether there are any normal cleanups on the stack.
420 bool hasNormalCleanups() const {
421 return InnermostNormalCleanup != stable_end();
424 /// Returns the innermost normal cleanup on the stack, or
425 /// stable_end() if there are no normal cleanups.
426 stable_iterator getInnermostNormalCleanup() const {
427 return InnermostNormalCleanup;
429 stable_iterator getInnermostActiveNormalCleanup() const; // CGException.h
431 /// Determines whether there are any EH cleanups on the stack.
432 bool hasEHCleanups() const {
433 return InnermostEHCleanup != stable_end();
436 /// Returns the innermost EH cleanup on the stack, or stable_end()
437 /// if there are no EH cleanups.
438 stable_iterator getInnermostEHCleanup() const {
439 return InnermostEHCleanup;
441 stable_iterator getInnermostActiveEHCleanup() const; // CGException.h
443 /// An unstable reference to a scope-stack depth. Invalidated by
444 /// pushes but not pops.
447 /// Returns an iterator pointing to the innermost EH scope.
448 iterator begin() const;
450 /// Returns an iterator pointing to the outermost EH scope.
451 iterator end() const;
453 /// Create a stable reference to the top of the EH stack. The
454 /// returned reference is valid until that scope is popped off the
456 stable_iterator stable_begin() const {
457 return stable_iterator(EndOfBuffer - StartOfData);
460 /// Create a stable reference to the bottom of the EH stack.
461 static stable_iterator stable_end() {
462 return stable_iterator(0);
465 /// Translates an iterator into a stable_iterator.
466 stable_iterator stabilize(iterator it) const;
468 /// Finds the nearest cleanup enclosing the given iterator.
469 /// Returns stable_iterator::invalid() if there are no such cleanups.
470 stable_iterator getEnclosingEHCleanup(iterator it) const;
472 /// Turn a stable reference to a scope depth into a unstable pointer
474 iterator find(stable_iterator save) const;
476 /// Removes the cleanup pointed to by the given stable_iterator.
477 void removeCleanup(stable_iterator save);
479 /// Add a branch fixup to the current cleanup scope.
480 BranchFixup &addBranchFixup() {
481 assert(hasNormalCleanups() && "adding fixup in scope without cleanups");
482 BranchFixups.push_back(BranchFixup());
483 return BranchFixups.back();
486 unsigned getNumBranchFixups() const { return BranchFixups.size(); }
487 BranchFixup &getBranchFixup(unsigned I) {
488 assert(I < getNumBranchFixups());
489 return BranchFixups[I];
492 /// Pops lazily-removed fixups from the end of the list. This
493 /// should only be called by procedures which have just popped a
494 /// cleanup or resolved one or more fixups.
495 void popNullFixups();
497 /// Clears the branch-fixups list. This should only be called by
498 /// ResolveAllBranchFixups.
499 void clearFixups() { BranchFixups.clear(); }
501 /// Gets the next EH destination index.
502 unsigned getNextEHDestIndex() { return NextEHDestIndex++; }
505 /// CodeGenFunction - This class organizes the per-function state that is used
506 /// while generating LLVM code.
507 class CodeGenFunction : public CodeGenTypeCache {
508 CodeGenFunction(const CodeGenFunction&); // DO NOT IMPLEMENT
509 void operator=(const CodeGenFunction&); // DO NOT IMPLEMENT
511 friend class CGCXXABI;
513 /// A jump destination is an abstract label, branching to which may
514 /// require a jump out through normal cleanups.
516 JumpDest() : Block(0), ScopeDepth(), Index(0) {}
517 JumpDest(llvm::BasicBlock *Block,
518 EHScopeStack::stable_iterator Depth,
520 : Block(Block), ScopeDepth(Depth), Index(Index) {}
522 bool isValid() const { return Block != 0; }
523 llvm::BasicBlock *getBlock() const { return Block; }
524 EHScopeStack::stable_iterator getScopeDepth() const { return ScopeDepth; }
525 unsigned getDestIndex() const { return Index; }
528 llvm::BasicBlock *Block;
529 EHScopeStack::stable_iterator ScopeDepth;
533 /// An unwind destination is an abstract label, branching to which
534 /// may require a jump out through EH cleanups.
536 UnwindDest() : Block(0), ScopeDepth(), Index(0) {}
537 UnwindDest(llvm::BasicBlock *Block,
538 EHScopeStack::stable_iterator Depth,
540 : Block(Block), ScopeDepth(Depth), Index(Index) {}
542 bool isValid() const { return Block != 0; }
543 llvm::BasicBlock *getBlock() const { return Block; }
544 EHScopeStack::stable_iterator getScopeDepth() const { return ScopeDepth; }
545 unsigned getDestIndex() const { return Index; }
548 llvm::BasicBlock *Block;
549 EHScopeStack::stable_iterator ScopeDepth;
553 CodeGenModule &CGM; // Per-module state.
554 const TargetInfo &Target;
556 typedef std::pair<llvm::Value *, llvm::Value *> ComplexPairTy;
559 /// CurFuncDecl - Holds the Decl for the current function or ObjC method.
560 /// This excludes BlockDecls.
561 const Decl *CurFuncDecl;
562 /// CurCodeDecl - This is the inner-most code context, which includes blocks.
563 const Decl *CurCodeDecl;
564 const CGFunctionInfo *CurFnInfo;
566 llvm::Function *CurFn;
568 /// CurGD - The GlobalDecl for the current function being compiled.
571 /// ReturnBlock - Unified return block.
572 JumpDest ReturnBlock;
574 /// ReturnValue - The temporary alloca to hold the return value. This is null
575 /// iff the function has no return value.
576 llvm::Value *ReturnValue;
578 /// RethrowBlock - Unified rethrow block.
579 UnwindDest RethrowBlock;
581 /// AllocaInsertPoint - This is an instruction in the entry block before which
582 /// we prefer to insert allocas.
583 llvm::AssertingVH<llvm::Instruction> AllocaInsertPt;
587 const CodeGen::CGBlockInfo *BlockInfo;
588 llvm::Value *BlockPointer;
590 /// \brief A mapping from NRVO variables to the flags used to indicate
591 /// when the NRVO has been applied to this variable.
592 llvm::DenseMap<const VarDecl *, llvm::Value *> NRVOFlags;
594 EHScopeStack EHStack;
596 /// i32s containing the indexes of the cleanup destinations.
597 llvm::AllocaInst *NormalCleanupDest;
598 llvm::AllocaInst *EHCleanupDest;
600 unsigned NextCleanupDestIndex;
602 /// The exception slot. All landing pads write the current
603 /// exception pointer into this alloca.
604 llvm::Value *ExceptionSlot;
606 /// Emits a landing pad for the current EH stack.
607 llvm::BasicBlock *EmitLandingPad();
609 llvm::BasicBlock *getInvokeDestImpl();
611 /// Set up the last cleaup that was pushed as a conditional
612 /// full-expression cleanup.
613 void initFullExprCleanup();
616 typename DominatingValue<T>::saved_type saveValueInCond(T value) {
617 return DominatingValue<T>::save(*this, value);
621 /// ObjCEHValueStack - Stack of Objective-C exception values, used for
623 llvm::SmallVector<llvm::Value*, 8> ObjCEHValueStack;
625 // A struct holding information about a finally block's IR
626 // generation. For now, doesn't actually hold anything.
630 FinallyInfo EnterFinallyBlock(const Stmt *Stmt,
631 llvm::Constant *BeginCatchFn,
632 llvm::Constant *EndCatchFn,
633 llvm::Constant *RethrowFn);
634 void ExitFinallyBlock(FinallyInfo &FinallyInfo);
636 /// pushFullExprCleanup - Push a cleanup to be run at the end of the
637 /// current full-expression. Safe against the possibility that
638 /// we're currently inside a conditionally-evaluated expression.
639 template <class T, class A0>
640 void pushFullExprCleanup(CleanupKind kind, A0 a0) {
641 // If we're not in a conditional branch, or if none of the
642 // arguments requires saving, then use the unconditional cleanup.
643 if (!isInConditionalBranch()) {
644 typedef EHScopeStack::UnconditionalCleanup1<T, A0> CleanupType;
645 return EHStack.pushCleanup<CleanupType>(kind, a0);
648 typename DominatingValue<A0>::saved_type a0_saved = saveValueInCond(a0);
650 typedef EHScopeStack::ConditionalCleanup1<T, A0> CleanupType;
651 EHStack.pushCleanup<CleanupType>(kind, a0_saved);
652 initFullExprCleanup();
655 /// pushFullExprCleanup - Push a cleanup to be run at the end of the
656 /// current full-expression. Safe against the possibility that
657 /// we're currently inside a conditionally-evaluated expression.
658 template <class T, class A0, class A1>
659 void pushFullExprCleanup(CleanupKind kind, A0 a0, A1 a1) {
660 // If we're not in a conditional branch, or if none of the
661 // arguments requires saving, then use the unconditional cleanup.
662 if (!isInConditionalBranch()) {
663 typedef EHScopeStack::UnconditionalCleanup2<T, A0, A1> CleanupType;
664 return EHStack.pushCleanup<CleanupType>(kind, a0, a1);
667 typename DominatingValue<A0>::saved_type a0_saved = saveValueInCond(a0);
668 typename DominatingValue<A1>::saved_type a1_saved = saveValueInCond(a1);
670 typedef EHScopeStack::ConditionalCleanup2<T, A0, A1> CleanupType;
671 EHStack.pushCleanup<CleanupType>(kind, a0_saved, a1_saved);
672 initFullExprCleanup();
675 /// PushDestructorCleanup - Push a cleanup to call the
676 /// complete-object destructor of an object of the given type at the
677 /// given address. Does nothing if T is not a C++ class type with a
678 /// non-trivial destructor.
679 void PushDestructorCleanup(QualType T, llvm::Value *Addr);
681 /// PushDestructorCleanup - Push a cleanup to call the
682 /// complete-object variant of the given destructor on the object at
683 /// the given address.
684 void PushDestructorCleanup(const CXXDestructorDecl *Dtor,
687 /// PopCleanupBlock - Will pop the cleanup entry on the stack and
688 /// process all branch fixups.
689 void PopCleanupBlock(bool FallThroughIsBranchThrough = false);
691 /// DeactivateCleanupBlock - Deactivates the given cleanup block.
692 /// The block cannot be reactivated. Pops it if it's the top of the
694 void DeactivateCleanupBlock(EHScopeStack::stable_iterator Cleanup);
696 /// ActivateCleanupBlock - Activates an initially-inactive cleanup.
697 /// Cannot be used to resurrect a deactivated cleanup.
698 void ActivateCleanupBlock(EHScopeStack::stable_iterator Cleanup);
700 /// \brief Enters a new scope for capturing cleanups, all of which
701 /// will be executed once the scope is exited.
702 class RunCleanupsScope {
703 CodeGenFunction& CGF;
704 EHScopeStack::stable_iterator CleanupStackDepth;
705 bool OldDidCallStackSave;
708 RunCleanupsScope(const RunCleanupsScope &); // DO NOT IMPLEMENT
709 RunCleanupsScope &operator=(const RunCleanupsScope &); // DO NOT IMPLEMENT
712 /// \brief Enter a new cleanup scope.
713 explicit RunCleanupsScope(CodeGenFunction &CGF)
714 : CGF(CGF), PerformCleanup(true)
716 CleanupStackDepth = CGF.EHStack.stable_begin();
717 OldDidCallStackSave = CGF.DidCallStackSave;
718 CGF.DidCallStackSave = false;
721 /// \brief Exit this cleanup scope, emitting any accumulated
723 ~RunCleanupsScope() {
724 if (PerformCleanup) {
725 CGF.DidCallStackSave = OldDidCallStackSave;
726 CGF.PopCleanupBlocks(CleanupStackDepth);
730 /// \brief Determine whether this scope requires any cleanups.
731 bool requiresCleanups() const {
732 return CGF.EHStack.stable_begin() != CleanupStackDepth;
735 /// \brief Force the emission of cleanups now, instead of waiting
736 /// until this object is destroyed.
737 void ForceCleanup() {
738 assert(PerformCleanup && "Already forced cleanup");
739 CGF.DidCallStackSave = OldDidCallStackSave;
740 CGF.PopCleanupBlocks(CleanupStackDepth);
741 PerformCleanup = false;
746 /// PopCleanupBlocks - Takes the old cleanup stack size and emits
747 /// the cleanup blocks that have been added.
748 void PopCleanupBlocks(EHScopeStack::stable_iterator OldCleanupStackSize);
750 void ResolveBranchFixups(llvm::BasicBlock *Target);
752 /// The given basic block lies in the current EH scope, but may be a
753 /// target of a potentially scope-crossing jump; get a stable handle
754 /// to which we can perform this jump later.
755 JumpDest getJumpDestInCurrentScope(llvm::BasicBlock *Target) {
756 return JumpDest(Target,
757 EHStack.getInnermostNormalCleanup(),
758 NextCleanupDestIndex++);
761 /// The given basic block lies in the current EH scope, but may be a
762 /// target of a potentially scope-crossing jump; get a stable handle
763 /// to which we can perform this jump later.
764 JumpDest getJumpDestInCurrentScope(llvm::StringRef Name = llvm::StringRef()) {
765 return getJumpDestInCurrentScope(createBasicBlock(Name));
768 /// EmitBranchThroughCleanup - Emit a branch from the current insert
769 /// block through the normal cleanup handling code (if any) and then
771 void EmitBranchThroughCleanup(JumpDest Dest);
773 /// isObviouslyBranchWithoutCleanups - Return true if a branch to the
774 /// specified destination obviously has no cleanups to run. 'false' is always
775 /// a conservatively correct answer for this method.
776 bool isObviouslyBranchWithoutCleanups(JumpDest Dest) const;
778 /// EmitBranchThroughEHCleanup - Emit a branch from the current
779 /// insert block through the EH cleanup handling code (if any) and
780 /// then on to \arg Dest.
781 void EmitBranchThroughEHCleanup(UnwindDest Dest);
783 /// getRethrowDest - Returns the unified outermost-scope rethrow
785 UnwindDest getRethrowDest();
787 /// An object to manage conditionally-evaluated expressions.
788 class ConditionalEvaluation {
789 llvm::BasicBlock *StartBB;
792 ConditionalEvaluation(CodeGenFunction &CGF)
793 : StartBB(CGF.Builder.GetInsertBlock()) {}
795 void begin(CodeGenFunction &CGF) {
796 assert(CGF.OutermostConditional != this);
797 if (!CGF.OutermostConditional)
798 CGF.OutermostConditional = this;
801 void end(CodeGenFunction &CGF) {
802 assert(CGF.OutermostConditional != 0);
803 if (CGF.OutermostConditional == this)
804 CGF.OutermostConditional = 0;
807 /// Returns a block which will be executed prior to each
808 /// evaluation of the conditional code.
809 llvm::BasicBlock *getStartingBlock() const {
814 /// isInConditionalBranch - Return true if we're currently emitting
815 /// one branch or the other of a conditional expression.
816 bool isInConditionalBranch() const { return OutermostConditional != 0; }
818 /// An RAII object to record that we're evaluating a statement
820 class StmtExprEvaluation {
821 CodeGenFunction &CGF;
823 /// We have to save the outermost conditional: cleanups in a
824 /// statement expression aren't conditional just because the
826 ConditionalEvaluation *SavedOutermostConditional;
829 StmtExprEvaluation(CodeGenFunction &CGF)
830 : CGF(CGF), SavedOutermostConditional(CGF.OutermostConditional) {
831 CGF.OutermostConditional = 0;
834 ~StmtExprEvaluation() {
835 CGF.OutermostConditional = SavedOutermostConditional;
836 CGF.EnsureInsertPoint();
840 /// An object which temporarily prevents a value from being
841 /// destroyed by aggressive peephole optimizations that assume that
842 /// all uses of a value have been realized in the IR.
843 class PeepholeProtection {
844 llvm::Instruction *Inst;
845 friend class CodeGenFunction;
848 PeepholeProtection() : Inst(0) {}
851 /// An RAII object to set (and then clear) a mapping for an OpaqueValueExpr.
852 class OpaqueValueMapping {
853 CodeGenFunction &CGF;
854 const OpaqueValueExpr *OpaqueValue;
856 CodeGenFunction::PeepholeProtection Protection;
859 static bool shouldBindAsLValue(const Expr *expr) {
860 return expr->isGLValue() || expr->getType()->isRecordType();
863 /// Build the opaque value mapping for the given conditional
864 /// operator if it's the GNU ?: extension. This is a common
865 /// enough pattern that the convenience operator is really
868 OpaqueValueMapping(CodeGenFunction &CGF,
869 const AbstractConditionalOperator *op) : CGF(CGF) {
870 if (isa<ConditionalOperator>(op)) {
876 const BinaryConditionalOperator *e = cast<BinaryConditionalOperator>(op);
877 init(e->getOpaqueValue(), e->getCommon());
880 OpaqueValueMapping(CodeGenFunction &CGF,
881 const OpaqueValueExpr *opaqueValue,
883 : CGF(CGF), OpaqueValue(opaqueValue), BoundLValue(true) {
884 assert(opaqueValue && "no opaque value expression!");
885 assert(shouldBindAsLValue(opaqueValue));
889 OpaqueValueMapping(CodeGenFunction &CGF,
890 const OpaqueValueExpr *opaqueValue,
892 : CGF(CGF), OpaqueValue(opaqueValue), BoundLValue(false) {
893 assert(opaqueValue && "no opaque value expression!");
894 assert(!shouldBindAsLValue(opaqueValue));
899 assert(OpaqueValue && "mapping already popped!");
904 ~OpaqueValueMapping() {
905 if (OpaqueValue) popImpl();
911 CGF.OpaqueLValues.erase(OpaqueValue);
913 CGF.OpaqueRValues.erase(OpaqueValue);
914 CGF.unprotectFromPeepholes(Protection);
918 void init(const OpaqueValueExpr *ov, const Expr *e) {
920 BoundLValue = shouldBindAsLValue(ov);
921 assert(BoundLValue == shouldBindAsLValue(e)
922 && "inconsistent expression value kinds!");
924 initLValue(CGF.EmitLValue(e));
926 initRValue(CGF.EmitAnyExpr(e));
929 void initLValue(const LValue &lv) {
930 CGF.OpaqueLValues.insert(std::make_pair(OpaqueValue, lv));
933 void initRValue(const RValue &rv) {
934 // Work around an extremely aggressive peephole optimization in
935 // EmitScalarConversion which assumes that all other uses of a
937 Protection = CGF.protectFromPeepholes(rv);
938 CGF.OpaqueRValues.insert(std::make_pair(OpaqueValue, rv));
942 /// getByrefValueFieldNumber - Given a declaration, returns the LLVM field
943 /// number that holds the value.
944 unsigned getByRefValueLLVMField(const ValueDecl *VD) const;
946 /// BuildBlockByrefAddress - Computes address location of the
947 /// variable which is declared as __block.
948 llvm::Value *BuildBlockByrefAddress(llvm::Value *BaseAddr,
951 CGDebugInfo *DebugInfo;
952 bool DisableDebugInfo;
954 /// IndirectBranch - The first time an indirect goto is seen we create a block
955 /// with an indirect branch. Every time we see the address of a label taken,
956 /// we add the label to the indirect goto. Every subsequent indirect goto is
957 /// codegen'd as a jump to the IndirectBranch's basic block.
958 llvm::IndirectBrInst *IndirectBranch;
960 /// LocalDeclMap - This keeps track of the LLVM allocas or globals for local C
962 typedef llvm::DenseMap<const Decl*, llvm::Value*> DeclMapTy;
963 DeclMapTy LocalDeclMap;
965 /// LabelMap - This keeps track of the LLVM basic block for each C label.
966 llvm::DenseMap<const LabelDecl*, JumpDest> LabelMap;
968 // BreakContinueStack - This keeps track of where break and continue
969 // statements should jump to.
970 struct BreakContinue {
971 BreakContinue(JumpDest Break, JumpDest Continue)
972 : BreakBlock(Break), ContinueBlock(Continue) {}
975 JumpDest ContinueBlock;
977 llvm::SmallVector<BreakContinue, 8> BreakContinueStack;
979 /// SwitchInsn - This is nearest current switch instruction. It is null if if
980 /// current context is not in a switch.
981 llvm::SwitchInst *SwitchInsn;
983 /// CaseRangeBlock - This block holds if condition check for last case
984 /// statement range in current switch instruction.
985 llvm::BasicBlock *CaseRangeBlock;
987 /// OpaqueLValues - Keeps track of the current set of opaque value
989 llvm::DenseMap<const OpaqueValueExpr *, LValue> OpaqueLValues;
990 llvm::DenseMap<const OpaqueValueExpr *, RValue> OpaqueRValues;
992 // VLASizeMap - This keeps track of the associated size for each VLA type.
993 // We track this by the size expression rather than the type itself because
994 // in certain situations, like a const qualifier applied to an VLA typedef,
995 // multiple VLA types can share the same size expression.
996 // FIXME: Maybe this could be a stack of maps that is pushed/popped as we
997 // enter/leave scopes.
998 llvm::DenseMap<const Expr*, llvm::Value*> VLASizeMap;
1000 /// DidCallStackSave - Whether llvm.stacksave has been called. Used to avoid
1001 /// calling llvm.stacksave for multiple VLAs in the same scope.
1002 bool DidCallStackSave;
1004 /// A block containing a single 'unreachable' instruction. Created
1005 /// lazily by getUnreachableBlock().
1006 llvm::BasicBlock *UnreachableBlock;
1008 /// CXXThisDecl - When generating code for a C++ member function,
1009 /// this will hold the implicit 'this' declaration.
1010 ImplicitParamDecl *CXXThisDecl;
1011 llvm::Value *CXXThisValue;
1013 /// CXXVTTDecl - When generating code for a base object constructor or
1014 /// base object destructor with virtual bases, this will hold the implicit
1016 ImplicitParamDecl *CXXVTTDecl;
1017 llvm::Value *CXXVTTValue;
1019 /// OutermostConditional - Points to the outermost active
1020 /// conditional control. This is used so that we know if a
1021 /// temporary should be destroyed conditionally.
1022 ConditionalEvaluation *OutermostConditional;
1025 /// ByrefValueInfoMap - For each __block variable, contains a pair of the LLVM
1026 /// type as well as the field number that contains the actual data.
1027 llvm::DenseMap<const ValueDecl *, std::pair<const llvm::Type *,
1028 unsigned> > ByRefValueInfo;
1030 llvm::BasicBlock *TerminateLandingPad;
1031 llvm::BasicBlock *TerminateHandler;
1032 llvm::BasicBlock *TrapBB;
1035 CodeGenFunction(CodeGenModule &cgm);
1037 CodeGenTypes &getTypes() const { return CGM.getTypes(); }
1038 ASTContext &getContext() const;
1039 CGDebugInfo *getDebugInfo() {
1040 if (DisableDebugInfo)
1044 void disableDebugInfo() { DisableDebugInfo = true; }
1045 void enableDebugInfo() { DisableDebugInfo = false; }
1048 const LangOptions &getLangOptions() const { return CGM.getLangOptions(); }
1050 /// Returns a pointer to the function's exception object slot, which
1051 /// is assigned in every landing pad.
1052 llvm::Value *getExceptionSlot();
1054 llvm::Value *getNormalCleanupDestSlot();
1055 llvm::Value *getEHCleanupDestSlot();
1057 llvm::BasicBlock *getUnreachableBlock() {
1058 if (!UnreachableBlock) {
1059 UnreachableBlock = createBasicBlock("unreachable");
1060 new llvm::UnreachableInst(getLLVMContext(), UnreachableBlock);
1062 return UnreachableBlock;
1065 llvm::BasicBlock *getInvokeDest() {
1066 if (!EHStack.requiresLandingPad()) return 0;
1067 return getInvokeDestImpl();
1070 llvm::LLVMContext &getLLVMContext() { return CGM.getLLVMContext(); }
1072 //===--------------------------------------------------------------------===//
1074 //===--------------------------------------------------------------------===//
1076 void GenerateObjCMethod(const ObjCMethodDecl *OMD);
1078 void StartObjCMethod(const ObjCMethodDecl *MD,
1079 const ObjCContainerDecl *CD);
1081 /// GenerateObjCGetter - Synthesize an Objective-C property getter function.
1082 void GenerateObjCGetter(ObjCImplementationDecl *IMP,
1083 const ObjCPropertyImplDecl *PID);
1084 void GenerateObjCGetterBody(ObjCIvarDecl *Ivar, bool IsAtomic, bool IsStrong);
1085 void GenerateObjCAtomicSetterBody(ObjCMethodDecl *OMD,
1086 ObjCIvarDecl *Ivar);
1088 void GenerateObjCCtorDtorMethod(ObjCImplementationDecl *IMP,
1089 ObjCMethodDecl *MD, bool ctor);
1091 /// GenerateObjCSetter - Synthesize an Objective-C property setter function
1092 /// for the given property.
1093 void GenerateObjCSetter(ObjCImplementationDecl *IMP,
1094 const ObjCPropertyImplDecl *PID);
1095 bool IndirectObjCSetterArg(const CGFunctionInfo &FI);
1096 bool IvarTypeWithAggrGCObjects(QualType Ty);
1098 //===--------------------------------------------------------------------===//
1100 //===--------------------------------------------------------------------===//
1102 llvm::Value *EmitBlockLiteral(const BlockExpr *);
1103 llvm::Constant *BuildDescriptorBlockDecl(const BlockExpr *,
1104 const CGBlockInfo &Info,
1105 const llvm::StructType *,
1106 llvm::Constant *BlockVarLayout);
1108 llvm::Function *GenerateBlockFunction(GlobalDecl GD,
1109 const CGBlockInfo &Info,
1110 const Decl *OuterFuncDecl,
1111 const DeclMapTy &ldm);
1113 llvm::Constant *GenerateCopyHelperFunction(const CGBlockInfo &blockInfo);
1114 llvm::Constant *GenerateDestroyHelperFunction(const CGBlockInfo &blockInfo);
1116 void BuildBlockRelease(llvm::Value *DeclPtr, BlockFieldFlags flags);
1118 class AutoVarEmission;
1120 void emitByrefStructureInit(const AutoVarEmission &emission);
1121 void enterByrefCleanup(const AutoVarEmission &emission);
1123 llvm::Value *LoadBlockStruct() {
1124 assert(BlockPointer && "no block pointer set!");
1125 return BlockPointer;
1128 void AllocateBlockCXXThisPointer(const CXXThisExpr *E);
1129 void AllocateBlockDecl(const BlockDeclRefExpr *E);
1130 llvm::Value *GetAddrOfBlockDecl(const BlockDeclRefExpr *E) {
1131 return GetAddrOfBlockDecl(E->getDecl(), E->isByRef());
1133 llvm::Value *GetAddrOfBlockDecl(const VarDecl *var, bool ByRef);
1134 const llvm::Type *BuildByRefType(const VarDecl *var);
1136 void GenerateCode(GlobalDecl GD, llvm::Function *Fn,
1137 const CGFunctionInfo &FnInfo);
1138 void StartFunction(GlobalDecl GD, QualType RetTy,
1140 const CGFunctionInfo &FnInfo,
1141 const FunctionArgList &Args,
1142 SourceLocation StartLoc);
1144 void EmitConstructorBody(FunctionArgList &Args);
1145 void EmitDestructorBody(FunctionArgList &Args);
1146 void EmitFunctionBody(FunctionArgList &Args);
1148 /// EmitReturnBlock - Emit the unified return block, trying to avoid its
1149 /// emission when possible.
1150 void EmitReturnBlock();
1152 /// FinishFunction - Complete IR generation of the current function. It is
1153 /// legal to call this function even if there is no current insertion point.
1154 void FinishFunction(SourceLocation EndLoc=SourceLocation());
1156 /// GenerateThunk - Generate a thunk for the given method.
1157 void GenerateThunk(llvm::Function *Fn, const CGFunctionInfo &FnInfo,
1158 GlobalDecl GD, const ThunkInfo &Thunk);
1160 void EmitCtorPrologue(const CXXConstructorDecl *CD, CXXCtorType Type,
1161 FunctionArgList &Args);
1163 /// InitializeVTablePointer - Initialize the vtable pointer of the given
1166 void InitializeVTablePointer(BaseSubobject Base,
1167 const CXXRecordDecl *NearestVBase,
1168 CharUnits OffsetFromNearestVBase,
1169 llvm::Constant *VTable,
1170 const CXXRecordDecl *VTableClass);
1172 typedef llvm::SmallPtrSet<const CXXRecordDecl *, 4> VisitedVirtualBasesSetTy;
1173 void InitializeVTablePointers(BaseSubobject Base,
1174 const CXXRecordDecl *NearestVBase,
1175 CharUnits OffsetFromNearestVBase,
1176 bool BaseIsNonVirtualPrimaryBase,
1177 llvm::Constant *VTable,
1178 const CXXRecordDecl *VTableClass,
1179 VisitedVirtualBasesSetTy& VBases);
1181 void InitializeVTablePointers(const CXXRecordDecl *ClassDecl);
1183 /// GetVTablePtr - Return the Value of the vtable pointer member pointed
1185 llvm::Value *GetVTablePtr(llvm::Value *This, const llvm::Type *Ty);
1187 /// EnterDtorCleanups - Enter the cleanups necessary to complete the
1188 /// given phase of destruction for a destructor. The end result
1189 /// should call destructors on members and base classes in reverse
1190 /// order of their construction.
1191 void EnterDtorCleanups(const CXXDestructorDecl *Dtor, CXXDtorType Type);
1193 /// ShouldInstrumentFunction - Return true if the current function should be
1194 /// instrumented with __cyg_profile_func_* calls
1195 bool ShouldInstrumentFunction();
1197 /// EmitFunctionInstrumentation - Emit LLVM code to call the specified
1198 /// instrumentation function with the current function and the call site, if
1199 /// function instrumentation is enabled.
1200 void EmitFunctionInstrumentation(const char *Fn);
1202 /// EmitMCountInstrumentation - Emit call to .mcount.
1203 void EmitMCountInstrumentation();
1205 /// EmitFunctionProlog - Emit the target specific LLVM code to load the
1206 /// arguments for the given function. This is also responsible for naming the
1207 /// LLVM function arguments.
1208 void EmitFunctionProlog(const CGFunctionInfo &FI,
1210 const FunctionArgList &Args);
1212 /// EmitFunctionEpilog - Emit the target specific LLVM code to return the
1213 /// given temporary.
1214 void EmitFunctionEpilog(const CGFunctionInfo &FI);
1216 /// EmitStartEHSpec - Emit the start of the exception spec.
1217 void EmitStartEHSpec(const Decl *D);
1219 /// EmitEndEHSpec - Emit the end of the exception spec.
1220 void EmitEndEHSpec(const Decl *D);
1222 /// getTerminateLandingPad - Return a landing pad that just calls terminate.
1223 llvm::BasicBlock *getTerminateLandingPad();
1225 /// getTerminateHandler - Return a handler (not a landing pad, just
1226 /// a catch handler) that just calls terminate. This is used when
1227 /// a terminate scope encloses a try.
1228 llvm::BasicBlock *getTerminateHandler();
1230 const llvm::Type *ConvertTypeForMem(QualType T);
1231 const llvm::Type *ConvertType(QualType T);
1232 const llvm::Type *ConvertType(const TypeDecl *T) {
1233 return ConvertType(getContext().getTypeDeclType(T));
1236 /// LoadObjCSelf - Load the value of self. This function is only valid while
1237 /// generating code for an Objective-C method.
1238 llvm::Value *LoadObjCSelf();
1240 /// TypeOfSelfObject - Return type of object that this self represents.
1241 QualType TypeOfSelfObject();
1243 /// hasAggregateLLVMType - Return true if the specified AST type will map into
1244 /// an aggregate LLVM type or is void.
1245 static bool hasAggregateLLVMType(QualType T);
1247 /// createBasicBlock - Create an LLVM basic block.
1248 llvm::BasicBlock *createBasicBlock(llvm::StringRef name = "",
1249 llvm::Function *parent = 0,
1250 llvm::BasicBlock *before = 0) {
1252 return llvm::BasicBlock::Create(getLLVMContext(), "", parent, before);
1254 return llvm::BasicBlock::Create(getLLVMContext(), name, parent, before);
1258 /// getBasicBlockForLabel - Return the LLVM basicblock that the specified
1260 JumpDest getJumpDestForLabel(const LabelDecl *S);
1262 /// SimplifyForwardingBlocks - If the given basic block is only a branch to
1263 /// another basic block, simplify it. This assumes that no other code could
1264 /// potentially reference the basic block.
1265 void SimplifyForwardingBlocks(llvm::BasicBlock *BB);
1267 /// EmitBlock - Emit the given block \arg BB and set it as the insert point,
1268 /// adding a fall-through branch from the current insert block if
1269 /// necessary. It is legal to call this function even if there is no current
1270 /// insertion point.
1272 /// IsFinished - If true, indicates that the caller has finished emitting
1273 /// branches to the given block and does not expect to emit code into it. This
1274 /// means the block can be ignored if it is unreachable.
1275 void EmitBlock(llvm::BasicBlock *BB, bool IsFinished=false);
1277 /// EmitBranch - Emit a branch to the specified basic block from the current
1278 /// insert block, taking care to avoid creation of branches from dummy
1279 /// blocks. It is legal to call this function even if there is no current
1280 /// insertion point.
1282 /// This function clears the current insertion point. The caller should follow
1283 /// calls to this function with calls to Emit*Block prior to generation new
1285 void EmitBranch(llvm::BasicBlock *Block);
1287 /// HaveInsertPoint - True if an insertion point is defined. If not, this
1288 /// indicates that the current code being emitted is unreachable.
1289 bool HaveInsertPoint() const {
1290 return Builder.GetInsertBlock() != 0;
1293 /// EnsureInsertPoint - Ensure that an insertion point is defined so that
1294 /// emitted IR has a place to go. Note that by definition, if this function
1295 /// creates a block then that block is unreachable; callers may do better to
1296 /// detect when no insertion point is defined and simply skip IR generation.
1297 void EnsureInsertPoint() {
1298 if (!HaveInsertPoint())
1299 EmitBlock(createBasicBlock());
1302 /// ErrorUnsupported - Print out an error that codegen doesn't support the
1303 /// specified stmt yet.
1304 void ErrorUnsupported(const Stmt *S, const char *Type,
1305 bool OmitOnError=false);
1307 //===--------------------------------------------------------------------===//
1309 //===--------------------------------------------------------------------===//
1311 LValue MakeAddrLValue(llvm::Value *V, QualType T, unsigned Alignment = 0) {
1312 return LValue::MakeAddr(V, T, Alignment, getContext(),
1313 CGM.getTBAAInfo(T));
1316 /// CreateTempAlloca - This creates a alloca and inserts it into the entry
1317 /// block. The caller is responsible for setting an appropriate alignment on
1319 llvm::AllocaInst *CreateTempAlloca(const llvm::Type *Ty,
1320 const llvm::Twine &Name = "tmp");
1322 /// InitTempAlloca - Provide an initial value for the given alloca.
1323 void InitTempAlloca(llvm::AllocaInst *Alloca, llvm::Value *Value);
1325 /// CreateIRTemp - Create a temporary IR object of the given type, with
1326 /// appropriate alignment. This routine should only be used when an temporary
1327 /// value needs to be stored into an alloca (for example, to avoid explicit
1328 /// PHI construction), but the type is the IR type, not the type appropriate
1329 /// for storing in memory.
1330 llvm::AllocaInst *CreateIRTemp(QualType T, const llvm::Twine &Name = "tmp");
1332 /// CreateMemTemp - Create a temporary memory object of the given type, with
1333 /// appropriate alignment.
1334 llvm::AllocaInst *CreateMemTemp(QualType T, const llvm::Twine &Name = "tmp");
1336 /// CreateAggTemp - Create a temporary memory object for the given
1338 AggValueSlot CreateAggTemp(QualType T, const llvm::Twine &Name = "tmp") {
1339 return AggValueSlot::forAddr(CreateMemTemp(T, Name), false, false);
1342 /// Emit a cast to void* in the appropriate address space.
1343 llvm::Value *EmitCastToVoidPtr(llvm::Value *value);
1345 /// EvaluateExprAsBool - Perform the usual unary conversions on the specified
1346 /// expression and compare the result against zero, returning an Int1Ty value.
1347 llvm::Value *EvaluateExprAsBool(const Expr *E);
1349 /// EmitIgnoredExpr - Emit an expression in a context which ignores the result.
1350 void EmitIgnoredExpr(const Expr *E);
1352 /// EmitAnyExpr - Emit code to compute the specified expression which can have
1353 /// any type. The result is returned as an RValue struct. If this is an
1354 /// aggregate expression, the aggloc/agglocvolatile arguments indicate where
1355 /// the result should be returned.
1357 /// \param IgnoreResult - True if the resulting value isn't used.
1358 RValue EmitAnyExpr(const Expr *E,
1359 AggValueSlot AggSlot = AggValueSlot::ignored(),
1360 bool IgnoreResult = false);
1362 // EmitVAListRef - Emit a "reference" to a va_list; this is either the address
1363 // or the value of the expression, depending on how va_list is defined.
1364 llvm::Value *EmitVAListRef(const Expr *E);
1366 /// EmitAnyExprToTemp - Similary to EmitAnyExpr(), however, the result will
1367 /// always be accessible even if no aggregate location is provided.
1368 RValue EmitAnyExprToTemp(const Expr *E);
1370 /// EmitAnyExprToMem - Emits the code necessary to evaluate an
1371 /// arbitrary expression into the given memory location.
1372 void EmitAnyExprToMem(const Expr *E, llvm::Value *Location,
1373 bool IsLocationVolatile,
1374 bool IsInitializer);
1376 /// EmitExprAsInit - Emits the code necessary to initialize a
1377 /// location in memory with the given initializer.
1378 void EmitExprAsInit(const Expr *init, const VarDecl *var,
1379 llvm::Value *loc, CharUnits alignment,
1380 bool capturedByInit);
1382 /// EmitAggregateCopy - Emit an aggrate copy.
1384 /// \param isVolatile - True iff either the source or the destination is
1386 void EmitAggregateCopy(llvm::Value *DestPtr, llvm::Value *SrcPtr,
1387 QualType EltTy, bool isVolatile=false);
1389 /// StartBlock - Start new block named N. If insert block is a dummy block
1391 void StartBlock(const char *N);
1393 /// GetAddrOfStaticLocalVar - Return the address of a static local variable.
1394 llvm::Constant *GetAddrOfStaticLocalVar(const VarDecl *BVD) {
1395 return cast<llvm::Constant>(GetAddrOfLocalVar(BVD));
1398 /// GetAddrOfLocalVar - Return the address of a local variable.
1399 llvm::Value *GetAddrOfLocalVar(const VarDecl *VD) {
1400 llvm::Value *Res = LocalDeclMap[VD];
1401 assert(Res && "Invalid argument to GetAddrOfLocalVar(), no decl!");
1405 /// getOpaqueLValueMapping - Given an opaque value expression (which
1406 /// must be mapped to an l-value), return its mapping.
1407 const LValue &getOpaqueLValueMapping(const OpaqueValueExpr *e) {
1408 assert(OpaqueValueMapping::shouldBindAsLValue(e));
1410 llvm::DenseMap<const OpaqueValueExpr*,LValue>::iterator
1411 it = OpaqueLValues.find(e);
1412 assert(it != OpaqueLValues.end() && "no mapping for opaque value!");
1416 /// getOpaqueRValueMapping - Given an opaque value expression (which
1417 /// must be mapped to an r-value), return its mapping.
1418 const RValue &getOpaqueRValueMapping(const OpaqueValueExpr *e) {
1419 assert(!OpaqueValueMapping::shouldBindAsLValue(e));
1421 llvm::DenseMap<const OpaqueValueExpr*,RValue>::iterator
1422 it = OpaqueRValues.find(e);
1423 assert(it != OpaqueRValues.end() && "no mapping for opaque value!");
1427 /// getAccessedFieldNo - Given an encoded value and a result number, return
1428 /// the input field number being accessed.
1429 static unsigned getAccessedFieldNo(unsigned Idx, const llvm::Constant *Elts);
1431 llvm::BlockAddress *GetAddrOfLabel(const LabelDecl *L);
1432 llvm::BasicBlock *GetIndirectGotoBlock();
1434 /// EmitNullInitialization - Generate code to set a value of the given type to
1435 /// null, If the type contains data member pointers, they will be initialized
1436 /// to -1 in accordance with the Itanium C++ ABI.
1437 void EmitNullInitialization(llvm::Value *DestPtr, QualType Ty);
1439 // EmitVAArg - Generate code to get an argument from the passed in pointer
1440 // and update it accordingly. The return value is a pointer to the argument.
1441 // FIXME: We should be able to get rid of this method and use the va_arg
1442 // instruction in LLVM instead once it works well enough.
1443 llvm::Value *EmitVAArg(llvm::Value *VAListAddr, QualType Ty);
1445 /// EmitVLASize - Generate code for any VLA size expressions that might occur
1446 /// in a variably modified type. If Ty is a VLA, will return the value that
1447 /// corresponds to the size in bytes of the VLA type. Will return 0 otherwise.
1449 /// This function can be called with a null (unreachable) insert point.
1450 llvm::Value *EmitVLASize(QualType Ty);
1452 // GetVLASize - Returns an LLVM value that corresponds to the size in bytes
1453 // of a variable length array type.
1454 llvm::Value *GetVLASize(const VariableArrayType *);
1456 /// LoadCXXThis - Load the value of 'this'. This function is only valid while
1457 /// generating code for an C++ member function.
1458 llvm::Value *LoadCXXThis() {
1459 assert(CXXThisValue && "no 'this' value for this function");
1460 return CXXThisValue;
1463 /// LoadCXXVTT - Load the VTT parameter to base constructors/destructors have
1465 llvm::Value *LoadCXXVTT() {
1466 assert(CXXVTTValue && "no VTT value for this function");
1470 /// GetAddressOfBaseOfCompleteClass - Convert the given pointer to a
1471 /// complete class to the given direct base.
1473 GetAddressOfDirectBaseInCompleteClass(llvm::Value *Value,
1474 const CXXRecordDecl *Derived,
1475 const CXXRecordDecl *Base,
1476 bool BaseIsVirtual);
1478 /// GetAddressOfBaseClass - This function will add the necessary delta to the
1479 /// load of 'this' and returns address of the base class.
1480 llvm::Value *GetAddressOfBaseClass(llvm::Value *Value,
1481 const CXXRecordDecl *Derived,
1482 CastExpr::path_const_iterator PathBegin,
1483 CastExpr::path_const_iterator PathEnd,
1484 bool NullCheckValue);
1486 llvm::Value *GetAddressOfDerivedClass(llvm::Value *Value,
1487 const CXXRecordDecl *Derived,
1488 CastExpr::path_const_iterator PathBegin,
1489 CastExpr::path_const_iterator PathEnd,
1490 bool NullCheckValue);
1492 llvm::Value *GetVirtualBaseClassOffset(llvm::Value *This,
1493 const CXXRecordDecl *ClassDecl,
1494 const CXXRecordDecl *BaseClassDecl);
1496 void EmitDelegateCXXConstructorCall(const CXXConstructorDecl *Ctor,
1497 CXXCtorType CtorType,
1498 const FunctionArgList &Args);
1499 // It's important not to confuse this and the previous function. Delegating
1500 // constructors are the C++0x feature. The constructor delegate optimization
1501 // is used to reduce duplication in the base and complete consturctors where
1502 // they are substantially the same.
1503 void EmitDelegatingCXXConstructorCall(const CXXConstructorDecl *Ctor,
1504 const FunctionArgList &Args);
1505 void EmitCXXConstructorCall(const CXXConstructorDecl *D, CXXCtorType Type,
1506 bool ForVirtualBase, llvm::Value *This,
1507 CallExpr::const_arg_iterator ArgBeg,
1508 CallExpr::const_arg_iterator ArgEnd);
1510 void EmitSynthesizedCXXCopyCtorCall(const CXXConstructorDecl *D,
1511 llvm::Value *This, llvm::Value *Src,
1512 CallExpr::const_arg_iterator ArgBeg,
1513 CallExpr::const_arg_iterator ArgEnd);
1515 void EmitCXXAggrConstructorCall(const CXXConstructorDecl *D,
1516 const ConstantArrayType *ArrayTy,
1517 llvm::Value *ArrayPtr,
1518 CallExpr::const_arg_iterator ArgBeg,
1519 CallExpr::const_arg_iterator ArgEnd,
1520 bool ZeroInitialization = false);
1522 void EmitCXXAggrConstructorCall(const CXXConstructorDecl *D,
1523 llvm::Value *NumElements,
1524 llvm::Value *ArrayPtr,
1525 CallExpr::const_arg_iterator ArgBeg,
1526 CallExpr::const_arg_iterator ArgEnd,
1527 bool ZeroInitialization = false);
1529 void EmitCXXAggrDestructorCall(const CXXDestructorDecl *D,
1530 const ArrayType *Array,
1533 void EmitCXXAggrDestructorCall(const CXXDestructorDecl *D,
1534 llvm::Value *NumElements,
1537 llvm::Function *GenerateCXXAggrDestructorHelper(const CXXDestructorDecl *D,
1538 const ArrayType *Array,
1541 void EmitCXXDestructorCall(const CXXDestructorDecl *D, CXXDtorType Type,
1542 bool ForVirtualBase, llvm::Value *This);
1544 void EmitNewArrayInitializer(const CXXNewExpr *E, llvm::Value *NewPtr,
1545 llvm::Value *NumElements);
1547 void EmitCXXTemporary(const CXXTemporary *Temporary, llvm::Value *Ptr);
1549 llvm::Value *EmitCXXNewExpr(const CXXNewExpr *E);
1550 void EmitCXXDeleteExpr(const CXXDeleteExpr *E);
1552 void EmitDeleteCall(const FunctionDecl *DeleteFD, llvm::Value *Ptr,
1555 llvm::Value* EmitCXXTypeidExpr(const CXXTypeidExpr *E);
1556 llvm::Value *EmitDynamicCast(llvm::Value *V, const CXXDynamicCastExpr *DCE);
1558 void EmitCheck(llvm::Value *, unsigned Size);
1560 llvm::Value *EmitScalarPrePostIncDec(const UnaryOperator *E, LValue LV,
1561 bool isInc, bool isPre);
1562 ComplexPairTy EmitComplexPrePostIncDec(const UnaryOperator *E, LValue LV,
1563 bool isInc, bool isPre);
1564 //===--------------------------------------------------------------------===//
1565 // Declaration Emission
1566 //===--------------------------------------------------------------------===//
1568 /// EmitDecl - Emit a declaration.
1570 /// This function can be called with a null (unreachable) insert point.
1571 void EmitDecl(const Decl &D);
1573 /// EmitVarDecl - Emit a local variable declaration.
1575 /// This function can be called with a null (unreachable) insert point.
1576 void EmitVarDecl(const VarDecl &D);
1578 typedef void SpecialInitFn(CodeGenFunction &Init, const VarDecl &D,
1579 llvm::Value *Address);
1581 /// EmitAutoVarDecl - Emit an auto variable declaration.
1583 /// This function can be called with a null (unreachable) insert point.
1584 void EmitAutoVarDecl(const VarDecl &D);
1586 class AutoVarEmission {
1587 friend class CodeGenFunction;
1589 const VarDecl *Variable;
1591 /// The alignment of the variable.
1592 CharUnits Alignment;
1594 /// The address of the alloca. Null if the variable was emitted
1595 /// as a global constant.
1596 llvm::Value *Address;
1598 llvm::Value *NRVOFlag;
1600 /// True if the variable is a __block variable.
1603 /// True if the variable is of aggregate type and has a constant
1605 bool IsConstantAggregate;
1608 AutoVarEmission(Invalid) : Variable(0) {}
1610 AutoVarEmission(const VarDecl &variable)
1611 : Variable(&variable), Address(0), NRVOFlag(0),
1612 IsByRef(false), IsConstantAggregate(false) {}
1614 bool wasEmittedAsGlobal() const { return Address == 0; }
1617 static AutoVarEmission invalid() { return AutoVarEmission(Invalid()); }
1619 /// Returns the address of the object within this declaration.
1620 /// Note that this does not chase the forwarding pointer for
1622 llvm::Value *getObjectAddress(CodeGenFunction &CGF) const {
1623 if (!IsByRef) return Address;
1625 return CGF.Builder.CreateStructGEP(Address,
1626 CGF.getByRefValueLLVMField(Variable),
1627 Variable->getNameAsString());
1630 AutoVarEmission EmitAutoVarAlloca(const VarDecl &var);
1631 void EmitAutoVarInit(const AutoVarEmission &emission);
1632 void EmitAutoVarCleanups(const AutoVarEmission &emission);
1634 void EmitStaticVarDecl(const VarDecl &D,
1635 llvm::GlobalValue::LinkageTypes Linkage);
1637 /// EmitParmDecl - Emit a ParmVarDecl or an ImplicitParamDecl.
1638 void EmitParmDecl(const VarDecl &D, llvm::Value *Arg, unsigned ArgNo);
1640 /// protectFromPeepholes - Protect a value that we're intending to
1641 /// store to the side, but which will probably be used later, from
1642 /// aggressive peepholing optimizations that might delete it.
1644 /// Pass the result to unprotectFromPeepholes to declare that
1645 /// protection is no longer required.
1647 /// There's no particular reason why this shouldn't apply to
1648 /// l-values, it's just that no existing peepholes work on pointers.
1649 PeepholeProtection protectFromPeepholes(RValue rvalue);
1650 void unprotectFromPeepholes(PeepholeProtection protection);
1652 //===--------------------------------------------------------------------===//
1653 // Statement Emission
1654 //===--------------------------------------------------------------------===//
1656 /// EmitStopPoint - Emit a debug stoppoint if we are emitting debug info.
1657 void EmitStopPoint(const Stmt *S);
1659 /// EmitStmt - Emit the code for the statement \arg S. It is legal to call
1660 /// this function even if there is no current insertion point.
1662 /// This function may clear the current insertion point; callers should use
1663 /// EnsureInsertPoint if they wish to subsequently generate code without first
1664 /// calling EmitBlock, EmitBranch, or EmitStmt.
1665 void EmitStmt(const Stmt *S);
1667 /// EmitSimpleStmt - Try to emit a "simple" statement which does not
1668 /// necessarily require an insertion point or debug information; typically
1669 /// because the statement amounts to a jump or a container of other
1672 /// \return True if the statement was handled.
1673 bool EmitSimpleStmt(const Stmt *S);
1675 RValue EmitCompoundStmt(const CompoundStmt &S, bool GetLast = false,
1676 AggValueSlot AVS = AggValueSlot::ignored());
1678 /// EmitLabel - Emit the block for the given label. It is legal to call this
1679 /// function even if there is no current insertion point.
1680 void EmitLabel(const LabelDecl *D); // helper for EmitLabelStmt.
1682 void EmitLabelStmt(const LabelStmt &S);
1683 void EmitGotoStmt(const GotoStmt &S);
1684 void EmitIndirectGotoStmt(const IndirectGotoStmt &S);
1685 void EmitIfStmt(const IfStmt &S);
1686 void EmitWhileStmt(const WhileStmt &S);
1687 void EmitDoStmt(const DoStmt &S);
1688 void EmitForStmt(const ForStmt &S);
1689 void EmitReturnStmt(const ReturnStmt &S);
1690 void EmitDeclStmt(const DeclStmt &S);
1691 void EmitBreakStmt(const BreakStmt &S);
1692 void EmitContinueStmt(const ContinueStmt &S);
1693 void EmitSwitchStmt(const SwitchStmt &S);
1694 void EmitDefaultStmt(const DefaultStmt &S);
1695 void EmitCaseStmt(const CaseStmt &S);
1696 void EmitCaseStmtRange(const CaseStmt &S);
1697 void EmitAsmStmt(const AsmStmt &S);
1699 void EmitObjCForCollectionStmt(const ObjCForCollectionStmt &S);
1700 void EmitObjCAtTryStmt(const ObjCAtTryStmt &S);
1701 void EmitObjCAtThrowStmt(const ObjCAtThrowStmt &S);
1702 void EmitObjCAtSynchronizedStmt(const ObjCAtSynchronizedStmt &S);
1704 llvm::Constant *getUnwindResumeOrRethrowFn();
1705 void EnterCXXTryStmt(const CXXTryStmt &S, bool IsFnTryBlock = false);
1706 void ExitCXXTryStmt(const CXXTryStmt &S, bool IsFnTryBlock = false);
1708 void EmitCXXTryStmt(const CXXTryStmt &S);
1709 void EmitCXXForRangeStmt(const CXXForRangeStmt &S);
1711 //===--------------------------------------------------------------------===//
1712 // LValue Expression Emission
1713 //===--------------------------------------------------------------------===//
1715 /// GetUndefRValue - Get an appropriate 'undef' rvalue for the given type.
1716 RValue GetUndefRValue(QualType Ty);
1718 /// EmitUnsupportedRValue - Emit a dummy r-value using the type of E
1719 /// and issue an ErrorUnsupported style diagnostic (using the
1721 RValue EmitUnsupportedRValue(const Expr *E,
1724 /// EmitUnsupportedLValue - Emit a dummy l-value using the type of E and issue
1725 /// an ErrorUnsupported style diagnostic (using the provided Name).
1726 LValue EmitUnsupportedLValue(const Expr *E,
1729 /// EmitLValue - Emit code to compute a designator that specifies the location
1730 /// of the expression.
1732 /// This can return one of two things: a simple address or a bitfield
1733 /// reference. In either case, the LLVM Value* in the LValue structure is
1734 /// guaranteed to be an LLVM pointer type.
1736 /// If this returns a bitfield reference, nothing about the pointee type of
1737 /// the LLVM value is known: For example, it may not be a pointer to an
1740 /// If this returns a normal address, and if the lvalue's C type is fixed
1741 /// size, this method guarantees that the returned pointer type will point to
1742 /// an LLVM type of the same size of the lvalue's type. If the lvalue has a
1743 /// variable length type, this is not possible.
1745 LValue EmitLValue(const Expr *E);
1747 /// EmitCheckedLValue - Same as EmitLValue but additionally we generate
1748 /// checking code to guard against undefined behavior. This is only
1749 /// suitable when we know that the address will be used to access the
1751 LValue EmitCheckedLValue(const Expr *E);
1753 /// EmitToMemory - Change a scalar value from its value
1754 /// representation to its in-memory representation.
1755 llvm::Value *EmitToMemory(llvm::Value *Value, QualType Ty);
1757 /// EmitFromMemory - Change a scalar value from its memory
1758 /// representation to its value representation.
1759 llvm::Value *EmitFromMemory(llvm::Value *Value, QualType Ty);
1761 /// EmitLoadOfScalar - Load a scalar value from an address, taking
1762 /// care to appropriately convert from the memory representation to
1763 /// the LLVM value representation.
1764 llvm::Value *EmitLoadOfScalar(llvm::Value *Addr, bool Volatile,
1765 unsigned Alignment, QualType Ty,
1766 llvm::MDNode *TBAAInfo = 0);
1768 /// EmitStoreOfScalar - Store a scalar value to an address, taking
1769 /// care to appropriately convert from the memory representation to
1770 /// the LLVM value representation.
1771 void EmitStoreOfScalar(llvm::Value *Value, llvm::Value *Addr,
1772 bool Volatile, unsigned Alignment, QualType Ty,
1773 llvm::MDNode *TBAAInfo = 0);
1775 /// EmitLoadOfLValue - Given an expression that represents a value lvalue,
1776 /// this method emits the address of the lvalue, then loads the result as an
1777 /// rvalue, returning the rvalue.
1778 RValue EmitLoadOfLValue(LValue V, QualType LVType);
1779 RValue EmitLoadOfExtVectorElementLValue(LValue V, QualType LVType);
1780 RValue EmitLoadOfBitfieldLValue(LValue LV, QualType ExprType);
1781 RValue EmitLoadOfPropertyRefLValue(LValue LV,
1782 ReturnValueSlot Return = ReturnValueSlot());
1784 /// EmitStoreThroughLValue - Store the specified rvalue into the specified
1785 /// lvalue, where both are guaranteed to the have the same type, and that type
1787 void EmitStoreThroughLValue(RValue Src, LValue Dst, QualType Ty);
1788 void EmitStoreThroughExtVectorComponentLValue(RValue Src, LValue Dst,
1790 void EmitStoreThroughPropertyRefLValue(RValue Src, LValue Dst);
1792 /// EmitStoreThroughLValue - Store Src into Dst with same constraints as
1793 /// EmitStoreThroughLValue.
1795 /// \param Result [out] - If non-null, this will be set to a Value* for the
1796 /// bit-field contents after the store, appropriate for use as the result of
1797 /// an assignment to the bit-field.
1798 void EmitStoreThroughBitfieldLValue(RValue Src, LValue Dst, QualType Ty,
1799 llvm::Value **Result=0);
1801 /// Emit an l-value for an assignment (simple or compound) of complex type.
1802 LValue EmitComplexAssignmentLValue(const BinaryOperator *E);
1803 LValue EmitComplexCompoundAssignmentLValue(const CompoundAssignOperator *E);
1805 // Note: only available for agg return types
1806 LValue EmitBinaryOperatorLValue(const BinaryOperator *E);
1807 LValue EmitCompoundAssignmentLValue(const CompoundAssignOperator *E);
1808 // Note: only available for agg return types
1809 LValue EmitCallExprLValue(const CallExpr *E);
1810 // Note: only available for agg return types
1811 LValue EmitVAArgExprLValue(const VAArgExpr *E);
1812 LValue EmitDeclRefLValue(const DeclRefExpr *E);
1813 LValue EmitStringLiteralLValue(const StringLiteral *E);
1814 LValue EmitObjCEncodeExprLValue(const ObjCEncodeExpr *E);
1815 LValue EmitPredefinedLValue(const PredefinedExpr *E);
1816 LValue EmitUnaryOpLValue(const UnaryOperator *E);
1817 LValue EmitArraySubscriptExpr(const ArraySubscriptExpr *E);
1818 LValue EmitExtVectorElementExpr(const ExtVectorElementExpr *E);
1819 LValue EmitMemberExpr(const MemberExpr *E);
1820 LValue EmitObjCIsaExpr(const ObjCIsaExpr *E);
1821 LValue EmitCompoundLiteralLValue(const CompoundLiteralExpr *E);
1822 LValue EmitConditionalOperatorLValue(const AbstractConditionalOperator *E);
1823 LValue EmitCastLValue(const CastExpr *E);
1824 LValue EmitNullInitializationLValue(const CXXScalarValueInitExpr *E);
1825 LValue EmitOpaqueValueLValue(const OpaqueValueExpr *e);
1827 llvm::Value *EmitIvarOffset(const ObjCInterfaceDecl *Interface,
1828 const ObjCIvarDecl *Ivar);
1829 LValue EmitLValueForAnonRecordField(llvm::Value* Base,
1830 const IndirectFieldDecl* Field,
1831 unsigned CVRQualifiers);
1832 LValue EmitLValueForField(llvm::Value* Base, const FieldDecl* Field,
1833 unsigned CVRQualifiers);
1835 /// EmitLValueForFieldInitialization - Like EmitLValueForField, except that
1836 /// if the Field is a reference, this will return the address of the reference
1837 /// and not the address of the value stored in the reference.
1838 LValue EmitLValueForFieldInitialization(llvm::Value* Base,
1839 const FieldDecl* Field,
1840 unsigned CVRQualifiers);
1842 LValue EmitLValueForIvar(QualType ObjectTy,
1843 llvm::Value* Base, const ObjCIvarDecl *Ivar,
1844 unsigned CVRQualifiers);
1846 LValue EmitLValueForBitfield(llvm::Value* Base, const FieldDecl* Field,
1847 unsigned CVRQualifiers);
1849 LValue EmitBlockDeclRefLValue(const BlockDeclRefExpr *E);
1851 LValue EmitCXXConstructLValue(const CXXConstructExpr *E);
1852 LValue EmitCXXBindTemporaryLValue(const CXXBindTemporaryExpr *E);
1853 LValue EmitExprWithCleanupsLValue(const ExprWithCleanups *E);
1854 LValue EmitCXXTypeidLValue(const CXXTypeidExpr *E);
1856 LValue EmitObjCMessageExprLValue(const ObjCMessageExpr *E);
1857 LValue EmitObjCIvarRefLValue(const ObjCIvarRefExpr *E);
1858 LValue EmitObjCPropertyRefLValue(const ObjCPropertyRefExpr *E);
1859 LValue EmitStmtExprLValue(const StmtExpr *E);
1860 LValue EmitPointerToDataMemberBinaryExpr(const BinaryOperator *E);
1861 LValue EmitObjCSelectorLValue(const ObjCSelectorExpr *E);
1862 void EmitDeclRefExprDbgValue(const DeclRefExpr *E, llvm::Constant *Init);
1864 //===--------------------------------------------------------------------===//
1865 // Scalar Expression Emission
1866 //===--------------------------------------------------------------------===//
1868 /// EmitCall - Generate a call of the given function, expecting the given
1869 /// result type, and using the given argument list which specifies both the
1870 /// LLVM arguments and the types they were derived from.
1872 /// \param TargetDecl - If given, the decl of the function in a direct call;
1873 /// used to set attributes on the call (noreturn, etc.).
1874 RValue EmitCall(const CGFunctionInfo &FnInfo,
1875 llvm::Value *Callee,
1876 ReturnValueSlot ReturnValue,
1877 const CallArgList &Args,
1878 const Decl *TargetDecl = 0,
1879 llvm::Instruction **callOrInvoke = 0);
1881 RValue EmitCall(QualType FnType, llvm::Value *Callee,
1882 ReturnValueSlot ReturnValue,
1883 CallExpr::const_arg_iterator ArgBeg,
1884 CallExpr::const_arg_iterator ArgEnd,
1885 const Decl *TargetDecl = 0);
1886 RValue EmitCallExpr(const CallExpr *E,
1887 ReturnValueSlot ReturnValue = ReturnValueSlot());
1889 llvm::CallSite EmitCallOrInvoke(llvm::Value *Callee,
1890 llvm::Value * const *ArgBegin,
1891 llvm::Value * const *ArgEnd,
1892 const llvm::Twine &Name = "");
1894 llvm::Value *BuildVirtualCall(const CXXMethodDecl *MD, llvm::Value *This,
1895 const llvm::Type *Ty);
1896 llvm::Value *BuildVirtualCall(const CXXDestructorDecl *DD, CXXDtorType Type,
1897 llvm::Value *This, const llvm::Type *Ty);
1898 llvm::Value *BuildAppleKextVirtualCall(const CXXMethodDecl *MD,
1899 NestedNameSpecifier *Qual,
1900 const llvm::Type *Ty);
1902 llvm::Value *BuildAppleKextVirtualDestructorCall(const CXXDestructorDecl *DD,
1904 const CXXRecordDecl *RD);
1906 RValue EmitCXXMemberCall(const CXXMethodDecl *MD,
1907 llvm::Value *Callee,
1908 ReturnValueSlot ReturnValue,
1911 CallExpr::const_arg_iterator ArgBeg,
1912 CallExpr::const_arg_iterator ArgEnd);
1913 RValue EmitCXXMemberCallExpr(const CXXMemberCallExpr *E,
1914 ReturnValueSlot ReturnValue);
1915 RValue EmitCXXMemberPointerCallExpr(const CXXMemberCallExpr *E,
1916 ReturnValueSlot ReturnValue);
1918 RValue EmitCXXOperatorMemberCallExpr(const CXXOperatorCallExpr *E,
1919 const CXXMethodDecl *MD,
1920 ReturnValueSlot ReturnValue);
1923 RValue EmitBuiltinExpr(const FunctionDecl *FD,
1924 unsigned BuiltinID, const CallExpr *E);
1926 RValue EmitBlockCallExpr(const CallExpr *E, ReturnValueSlot ReturnValue);
1928 /// EmitTargetBuiltinExpr - Emit the given builtin call. Returns 0 if the call
1929 /// is unhandled by the current target.
1930 llvm::Value *EmitTargetBuiltinExpr(unsigned BuiltinID, const CallExpr *E);
1932 llvm::Value *EmitARMBuiltinExpr(unsigned BuiltinID, const CallExpr *E);
1933 llvm::Value *EmitNeonCall(llvm::Function *F,
1934 llvm::SmallVectorImpl<llvm::Value*> &O,
1936 unsigned shift = 0, bool rightshift = false);
1937 llvm::Value *EmitNeonSplat(llvm::Value *V, llvm::Constant *Idx);
1938 llvm::Value *EmitNeonShiftVector(llvm::Value *V, const llvm::Type *Ty,
1939 bool negateForRightShift);
1941 llvm::Value *BuildVector(const llvm::SmallVectorImpl<llvm::Value*> &Ops);
1942 llvm::Value *EmitX86BuiltinExpr(unsigned BuiltinID, const CallExpr *E);
1943 llvm::Value *EmitPPCBuiltinExpr(unsigned BuiltinID, const CallExpr *E);
1945 llvm::Value *EmitObjCProtocolExpr(const ObjCProtocolExpr *E);
1946 llvm::Value *EmitObjCStringLiteral(const ObjCStringLiteral *E);
1947 llvm::Value *EmitObjCSelectorExpr(const ObjCSelectorExpr *E);
1948 RValue EmitObjCMessageExpr(const ObjCMessageExpr *E,
1949 ReturnValueSlot Return = ReturnValueSlot());
1951 /// EmitReferenceBindingToExpr - Emits a reference binding to the passed in
1952 /// expression. Will emit a temporary variable if E is not an LValue.
1953 RValue EmitReferenceBindingToExpr(const Expr* E,
1954 const NamedDecl *InitializedDecl);
1956 //===--------------------------------------------------------------------===//
1957 // Expression Emission
1958 //===--------------------------------------------------------------------===//
1960 // Expressions are broken into three classes: scalar, complex, aggregate.
1962 /// EmitScalarExpr - Emit the computation of the specified expression of LLVM
1963 /// scalar type, returning the result.
1964 llvm::Value *EmitScalarExpr(const Expr *E , bool IgnoreResultAssign = false);
1966 /// EmitScalarConversion - Emit a conversion from the specified type to the
1967 /// specified destination type, both of which are LLVM scalar types.
1968 llvm::Value *EmitScalarConversion(llvm::Value *Src, QualType SrcTy,
1971 /// EmitComplexToScalarConversion - Emit a conversion from the specified
1972 /// complex type to the specified destination type, where the destination type
1973 /// is an LLVM scalar type.
1974 llvm::Value *EmitComplexToScalarConversion(ComplexPairTy Src, QualType SrcTy,
1978 /// EmitAggExpr - Emit the computation of the specified expression
1979 /// of aggregate type. The result is computed into the given slot,
1980 /// which may be null to indicate that the value is not needed.
1981 void EmitAggExpr(const Expr *E, AggValueSlot AS, bool IgnoreResult = false);
1983 /// EmitAggExprToLValue - Emit the computation of the specified expression of
1984 /// aggregate type into a temporary LValue.
1985 LValue EmitAggExprToLValue(const Expr *E);
1987 /// EmitGCMemmoveCollectable - Emit special API for structs with object
1989 void EmitGCMemmoveCollectable(llvm::Value *DestPtr, llvm::Value *SrcPtr,
1992 /// EmitComplexExpr - Emit the computation of the specified expression of
1993 /// complex type, returning the result.
1994 ComplexPairTy EmitComplexExpr(const Expr *E,
1995 bool IgnoreReal = false,
1996 bool IgnoreImag = false);
1998 /// EmitComplexExprIntoAddr - Emit the computation of the specified expression
1999 /// of complex type, storing into the specified Value*.
2000 void EmitComplexExprIntoAddr(const Expr *E, llvm::Value *DestAddr,
2001 bool DestIsVolatile);
2003 /// StoreComplexToAddr - Store a complex number into the specified address.
2004 void StoreComplexToAddr(ComplexPairTy V, llvm::Value *DestAddr,
2005 bool DestIsVolatile);
2006 /// LoadComplexFromAddr - Load a complex number from the specified address.
2007 ComplexPairTy LoadComplexFromAddr(llvm::Value *SrcAddr, bool SrcIsVolatile);
2009 /// CreateStaticVarDecl - Create a zero-initialized LLVM global for
2010 /// a static local variable.
2011 llvm::GlobalVariable *CreateStaticVarDecl(const VarDecl &D,
2012 const char *Separator,
2013 llvm::GlobalValue::LinkageTypes Linkage);
2015 /// AddInitializerToStaticVarDecl - Add the initializer for 'D' to the
2016 /// global variable that has already been created for it. If the initializer
2017 /// has a different type than GV does, this may free GV and return a different
2018 /// one. Otherwise it just returns GV.
2019 llvm::GlobalVariable *
2020 AddInitializerToStaticVarDecl(const VarDecl &D,
2021 llvm::GlobalVariable *GV);
2024 /// EmitCXXGlobalVarDeclInit - Create the initializer for a C++
2025 /// variable with global storage.
2026 void EmitCXXGlobalVarDeclInit(const VarDecl &D, llvm::Constant *DeclPtr);
2028 /// EmitCXXGlobalDtorRegistration - Emits a call to register the global ptr
2029 /// with the C++ runtime so that its destructor will be called at exit.
2030 void EmitCXXGlobalDtorRegistration(llvm::Constant *DtorFn,
2031 llvm::Constant *DeclPtr);
2033 /// Emit code in this function to perform a guarded variable
2034 /// initialization. Guarded initializations are used when it's not
2035 /// possible to prove that an initialization will be done exactly
2036 /// once, e.g. with a static local variable or a static data member
2037 /// of a class template.
2038 void EmitCXXGuardedInit(const VarDecl &D, llvm::GlobalVariable *DeclPtr);
2040 /// GenerateCXXGlobalInitFunc - Generates code for initializing global
2042 void GenerateCXXGlobalInitFunc(llvm::Function *Fn,
2043 llvm::Constant **Decls,
2046 /// GenerateCXXGlobalDtorFunc - Generates code for destroying global
2048 void GenerateCXXGlobalDtorFunc(llvm::Function *Fn,
2049 const std::vector<std::pair<llvm::WeakVH,
2050 llvm::Constant*> > &DtorsAndObjects);
2052 void GenerateCXXGlobalVarDeclInitFunc(llvm::Function *Fn,
2054 llvm::GlobalVariable *Addr);
2056 void EmitCXXConstructExpr(const CXXConstructExpr *E, AggValueSlot Dest);
2058 void EmitSynthesizedCXXCopyCtor(llvm::Value *Dest, llvm::Value *Src,
2061 RValue EmitExprWithCleanups(const ExprWithCleanups *E,
2062 AggValueSlot Slot =AggValueSlot::ignored());
2064 void EmitCXXThrowExpr(const CXXThrowExpr *E);
2066 //===--------------------------------------------------------------------===//
2068 //===--------------------------------------------------------------------===//
2070 /// ContainsLabel - Return true if the statement contains a label in it. If
2071 /// this statement is not executed normally, it not containing a label means
2072 /// that we can just remove the code.
2073 static bool ContainsLabel(const Stmt *S, bool IgnoreCaseStmts = false);
2075 /// containsBreak - Return true if the statement contains a break out of it.
2076 /// If the statement (recursively) contains a switch or loop with a break
2077 /// inside of it, this is fine.
2078 static bool containsBreak(const Stmt *S);
2080 /// ConstantFoldsToSimpleInteger - If the specified expression does not fold
2081 /// to a constant, or if it does but contains a label, return false. If it
2082 /// constant folds return true and set the boolean result in Result.
2083 bool ConstantFoldsToSimpleInteger(const Expr *Cond, bool &Result);
2085 /// ConstantFoldsToSimpleInteger - If the specified expression does not fold
2086 /// to a constant, or if it does but contains a label, return false. If it
2087 /// constant folds return true and set the folded value.
2088 bool ConstantFoldsToSimpleInteger(const Expr *Cond, llvm::APInt &Result);
2090 /// EmitBranchOnBoolExpr - Emit a branch on a boolean condition (e.g. for an
2091 /// if statement) to the specified blocks. Based on the condition, this might
2092 /// try to simplify the codegen of the conditional based on the branch.
2093 void EmitBranchOnBoolExpr(const Expr *Cond, llvm::BasicBlock *TrueBlock,
2094 llvm::BasicBlock *FalseBlock);
2096 /// getTrapBB - Create a basic block that will call the trap intrinsic. We'll
2097 /// generate a branch around the created basic block as necessary.
2098 llvm::BasicBlock *getTrapBB();
2100 /// EmitCallArg - Emit a single call argument.
2101 void EmitCallArg(CallArgList &args, const Expr *E, QualType ArgType);
2103 /// EmitDelegateCallArg - We are performing a delegate call; that
2104 /// is, the current function is delegating to another one. Produce
2105 /// a r-value suitable for passing the given parameter.
2106 void EmitDelegateCallArg(CallArgList &args, const VarDecl *param);
2109 void EmitReturnOfRValue(RValue RV, QualType Ty);
2111 /// ExpandTypeFromArgs - Reconstruct a structure of type \arg Ty
2112 /// from function arguments into \arg Dst. See ABIArgInfo::Expand.
2114 /// \param AI - The first function argument of the expansion.
2115 /// \return The argument following the last expanded function
2117 llvm::Function::arg_iterator
2118 ExpandTypeFromArgs(QualType Ty, LValue Dst,
2119 llvm::Function::arg_iterator AI);
2121 /// ExpandTypeToArgs - Expand an RValue \arg Src, with the LLVM type for \arg
2122 /// Ty, into individual arguments on the provided vector \arg Args. See
2123 /// ABIArgInfo::Expand.
2124 void ExpandTypeToArgs(QualType Ty, RValue Src,
2125 llvm::SmallVector<llvm::Value*, 16> &Args);
2127 llvm::Value* EmitAsmInput(const AsmStmt &S,
2128 const TargetInfo::ConstraintInfo &Info,
2129 const Expr *InputExpr, std::string &ConstraintStr);
2131 llvm::Value* EmitAsmInputLValue(const AsmStmt &S,
2132 const TargetInfo::ConstraintInfo &Info,
2133 LValue InputValue, QualType InputType,
2134 std::string &ConstraintStr);
2136 /// EmitCallArgs - Emit call arguments for a function.
2137 /// The CallArgTypeInfo parameter is used for iterating over the known
2138 /// argument types of the function being called.
2139 template<typename T>
2140 void EmitCallArgs(CallArgList& Args, const T* CallArgTypeInfo,
2141 CallExpr::const_arg_iterator ArgBeg,
2142 CallExpr::const_arg_iterator ArgEnd) {
2143 CallExpr::const_arg_iterator Arg = ArgBeg;
2145 // First, use the argument types that the type info knows about
2146 if (CallArgTypeInfo) {
2147 for (typename T::arg_type_iterator I = CallArgTypeInfo->arg_type_begin(),
2148 E = CallArgTypeInfo->arg_type_end(); I != E; ++I, ++Arg) {
2149 assert(Arg != ArgEnd && "Running over edge of argument list!");
2150 QualType ArgType = *I;
2152 QualType ActualArgType = Arg->getType();
2153 if (ArgType->isPointerType() && ActualArgType->isPointerType()) {
2154 QualType ActualBaseType =
2155 ActualArgType->getAs<PointerType>()->getPointeeType();
2156 QualType ArgBaseType =
2157 ArgType->getAs<PointerType>()->getPointeeType();
2158 if (ArgBaseType->isVariableArrayType()) {
2159 if (const VariableArrayType *VAT =
2160 getContext().getAsVariableArrayType(ActualBaseType)) {
2161 if (!VAT->getSizeExpr())
2162 ActualArgType = ArgType;
2166 assert(getContext().getCanonicalType(ArgType.getNonReferenceType()).
2168 getContext().getCanonicalType(ActualArgType).getTypePtr() &&
2169 "type mismatch in call argument!");
2171 EmitCallArg(Args, *Arg, ArgType);
2174 // Either we've emitted all the call args, or we have a call to a
2175 // variadic function.
2176 assert((Arg == ArgEnd || CallArgTypeInfo->isVariadic()) &&
2177 "Extra arguments in non-variadic function!");
2181 // If we still have any arguments, emit them using the type of the argument.
2182 for (; Arg != ArgEnd; ++Arg)
2183 EmitCallArg(Args, *Arg, Arg->getType());
2186 const TargetCodeGenInfo &getTargetHooks() const {
2187 return CGM.getTargetCodeGenInfo();
2190 void EmitDeclMetadata();
2192 CodeGenModule::ByrefHelpers *
2193 buildByrefHelpers(const llvm::StructType &byrefType,
2194 const AutoVarEmission &emission);
2197 /// Helper class with most of the code for saving a value for a
2198 /// conditional expression cleanup.
2199 struct DominatingLLVMValue {
2200 typedef llvm::PointerIntPair<llvm::Value*, 1, bool> saved_type;
2202 /// Answer whether the given value needs extra work to be saved.
2203 static bool needsSaving(llvm::Value *value) {
2204 // If it's not an instruction, we don't need to save.
2205 if (!isa<llvm::Instruction>(value)) return false;
2207 // If it's an instruction in the entry block, we don't need to save.
2208 llvm::BasicBlock *block = cast<llvm::Instruction>(value)->getParent();
2209 return (block != &block->getParent()->getEntryBlock());
2212 /// Try to save the given value.
2213 static saved_type save(CodeGenFunction &CGF, llvm::Value *value) {
2214 if (!needsSaving(value)) return saved_type(value, false);
2216 // Otherwise we need an alloca.
2217 llvm::Value *alloca =
2218 CGF.CreateTempAlloca(value->getType(), "cond-cleanup.save");
2219 CGF.Builder.CreateStore(value, alloca);
2221 return saved_type(alloca, true);
2224 static llvm::Value *restore(CodeGenFunction &CGF, saved_type value) {
2225 if (!value.getInt()) return value.getPointer();
2226 return CGF.Builder.CreateLoad(value.getPointer());
2230 /// A partial specialization of DominatingValue for llvm::Values that
2231 /// might be llvm::Instructions.
2232 template <class T> struct DominatingPointer<T,true> : DominatingLLVMValue {
2234 static type restore(CodeGenFunction &CGF, saved_type value) {
2235 return static_cast<T*>(DominatingLLVMValue::restore(CGF, value));
2239 /// A specialization of DominatingValue for RValue.
2240 template <> struct DominatingValue<RValue> {
2241 typedef RValue type;
2243 enum Kind { ScalarLiteral, ScalarAddress, AggregateLiteral,
2244 AggregateAddress, ComplexAddress };
2248 saved_type(llvm::Value *v, Kind k) : Value(v), K(k) {}
2251 static bool needsSaving(RValue value);
2252 static saved_type save(CodeGenFunction &CGF, RValue value);
2253 RValue restore(CodeGenFunction &CGF);
2255 // implementations in CGExprCXX.cpp
2258 static bool needsSaving(type value) {
2259 return saved_type::needsSaving(value);
2261 static saved_type save(CodeGenFunction &CGF, type value) {
2262 return saved_type::save(CGF, value);
2264 static type restore(CodeGenFunction &CGF, saved_type value) {
2265 return value.restore(CGF);
2269 } // end namespace CodeGen
2270 } // end namespace clang