clang 20.0.0git
CGBlocks.cpp
Go to the documentation of this file.
1//===--- CGBlocks.cpp - Emit LLVM Code for declarations ---------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This contains code to emit blocks.
10//
11//===----------------------------------------------------------------------===//
12
13#include "CGBlocks.h"
14#include "CGCXXABI.h"
15#include "CGDebugInfo.h"
16#include "CGObjCRuntime.h"
17#include "CGOpenCLRuntime.h"
18#include "CodeGenFunction.h"
19#include "CodeGenModule.h"
20#include "ConstantEmitter.h"
21#include "TargetInfo.h"
22#include "clang/AST/Attr.h"
23#include "clang/AST/DeclObjC.h"
25#include "llvm/IR/DataLayout.h"
26#include "llvm/IR/Module.h"
27#include "llvm/Support/ScopedPrinter.h"
28#include <algorithm>
29#include <cstdio>
30
31using namespace clang;
32using namespace CodeGen;
33
34CGBlockInfo::CGBlockInfo(const BlockDecl *block, StringRef name)
35 : Name(name), CXXThisIndex(0), CanBeGlobal(false), NeedsCopyDispose(false),
36 NoEscape(false), HasCXXObject(false), UsesStret(false),
37 HasCapturedVariableLayout(false), CapturesNonExternalType(false),
38 LocalAddress(RawAddress::invalid()), StructureType(nullptr),
39 Block(block) {
40
41 // Skip asm prefix, if any. 'name' is usually taken directly from
42 // the mangled name of the enclosing function.
43 if (!name.empty() && name[0] == '\01')
44 name = name.substr(1);
45}
46
47// Anchor the vtable to this translation unit.
49
50/// Build the given block as a global block.
51static llvm::Constant *buildGlobalBlock(CodeGenModule &CGM,
52 const CGBlockInfo &blockInfo,
53 llvm::Constant *blockFn);
54
55/// Build the helper function to copy a block.
56static llvm::Constant *buildCopyHelper(CodeGenModule &CGM,
57 const CGBlockInfo &blockInfo) {
58 return CodeGenFunction(CGM).GenerateCopyHelperFunction(blockInfo);
59}
60
61/// Build the helper function to dispose of a block.
62static llvm::Constant *buildDisposeHelper(CodeGenModule &CGM,
63 const CGBlockInfo &blockInfo) {
65}
66
67namespace {
68
69enum class CaptureStrKind {
70 // String for the copy helper.
71 CopyHelper,
72 // String for the dispose helper.
73 DisposeHelper,
74 // Merge the strings for the copy helper and dispose helper.
75 Merged
76};
77
78} // end anonymous namespace
79
80static std::string getBlockCaptureStr(const CGBlockInfo::Capture &Cap,
81 CaptureStrKind StrKind,
82 CharUnits BlockAlignment,
83 CodeGenModule &CGM);
84
85static std::string getBlockDescriptorName(const CGBlockInfo &BlockInfo,
86 CodeGenModule &CGM) {
87 std::string Name = "__block_descriptor_";
88 Name += llvm::to_string(BlockInfo.BlockSize.getQuantity()) + "_";
89
90 if (BlockInfo.NeedsCopyDispose) {
91 if (CGM.getLangOpts().Exceptions)
92 Name += "e";
93 if (CGM.getCodeGenOpts().ObjCAutoRefCountExceptions)
94 Name += "a";
95 Name += llvm::to_string(BlockInfo.BlockAlign.getQuantity()) + "_";
96
97 for (auto &Cap : BlockInfo.SortedCaptures) {
98 if (Cap.isConstantOrTrivial())
99 continue;
100
101 Name += llvm::to_string(Cap.getOffset().getQuantity());
102
103 if (Cap.CopyKind == Cap.DisposeKind) {
104 // If CopyKind and DisposeKind are the same, merge the capture
105 // information.
106 assert(Cap.CopyKind != BlockCaptureEntityKind::None &&
107 "shouldn't see BlockCaptureManagedEntity that is None");
108 Name += getBlockCaptureStr(Cap, CaptureStrKind::Merged,
109 BlockInfo.BlockAlign, CGM);
110 } else {
111 // If CopyKind and DisposeKind are not the same, which can happen when
112 // either Kind is None or the captured object is a __strong block,
113 // concatenate the copy and dispose strings.
114 Name += getBlockCaptureStr(Cap, CaptureStrKind::CopyHelper,
115 BlockInfo.BlockAlign, CGM);
116 Name += getBlockCaptureStr(Cap, CaptureStrKind::DisposeHelper,
117 BlockInfo.BlockAlign, CGM);
118 }
119 }
120 Name += "_";
121 }
122
123 std::string TypeAtEncoding;
124
125 if (!CGM.getCodeGenOpts().DisableBlockSignatureString) {
126 TypeAtEncoding =
128 /// Replace occurrences of '@' with '\1'. '@' is reserved on ELF platforms
129 /// as a separator between symbol name and symbol version.
130 std::replace(TypeAtEncoding.begin(), TypeAtEncoding.end(), '@', '\1');
131 }
132 Name += "e" + llvm::to_string(TypeAtEncoding.size()) + "_" + TypeAtEncoding;
133 Name += "l" + CGM.getObjCRuntime().getRCBlockLayoutStr(CGM, BlockInfo);
134 return Name;
135}
136
137/// buildBlockDescriptor - Build the block descriptor meta-data for a block.
138/// buildBlockDescriptor is accessed from 5th field of the Block_literal
139/// meta-data and contains stationary information about the block literal.
140/// Its definition will have 4 (or optionally 6) words.
141/// \code
142/// struct Block_descriptor {
143/// unsigned long reserved;
144/// unsigned long size; // size of Block_literal metadata in bytes.
145/// void *copy_func_helper_decl; // optional copy helper.
146/// void *destroy_func_decl; // optional destructor helper.
147/// void *block_method_encoding_address; // @encode for block literal signature.
148/// void *block_layout_info; // encoding of captured block variables.
149/// };
150/// \endcode
151static llvm::Constant *buildBlockDescriptor(CodeGenModule &CGM,
152 const CGBlockInfo &blockInfo) {
153 ASTContext &C = CGM.getContext();
154
155 llvm::IntegerType *ulong =
156 cast<llvm::IntegerType>(CGM.getTypes().ConvertType(C.UnsignedLongTy));
157 llvm::PointerType *i8p = nullptr;
158 if (CGM.getLangOpts().OpenCL)
159 i8p = llvm::PointerType::get(
160 CGM.getLLVMContext(), C.getTargetAddressSpace(LangAS::opencl_constant));
161 else
162 i8p = CGM.VoidPtrTy;
163
164 std::string descName;
165
166 // If an equivalent block descriptor global variable exists, return it.
167 if (C.getLangOpts().ObjC &&
168 CGM.getLangOpts().getGC() == LangOptions::NonGC) {
169 descName = getBlockDescriptorName(blockInfo, CGM);
170 if (llvm::GlobalValue *desc = CGM.getModule().getNamedValue(descName))
171 return desc;
172 }
173
174 // If there isn't an equivalent block descriptor global variable, create a new
175 // one.
176 ConstantInitBuilder builder(CGM);
177 auto elements = builder.beginStruct();
178
179 // reserved
180 elements.addInt(ulong, 0);
181
182 // Size
183 // FIXME: What is the right way to say this doesn't fit? We should give
184 // a user diagnostic in that case. Better fix would be to change the
185 // API to size_t.
186 elements.addInt(ulong, blockInfo.BlockSize.getQuantity());
187
188 // Optional copy/dispose helpers.
189 bool hasInternalHelper = false;
190 if (blockInfo.NeedsCopyDispose) {
191 // copy_func_helper_decl
192 llvm::Constant *copyHelper = buildCopyHelper(CGM, blockInfo);
193 elements.add(copyHelper);
194
195 // destroy_func_decl
196 llvm::Constant *disposeHelper = buildDisposeHelper(CGM, blockInfo);
197 elements.add(disposeHelper);
198
199 if (cast<llvm::Function>(copyHelper->stripPointerCasts())
200 ->hasInternalLinkage() ||
201 cast<llvm::Function>(disposeHelper->stripPointerCasts())
202 ->hasInternalLinkage())
203 hasInternalHelper = true;
204 }
205
206 // Signature. Mandatory ObjC-style method descriptor @encode sequence.
207 if (CGM.getCodeGenOpts().DisableBlockSignatureString) {
208 elements.addNullPointer(i8p);
209 } else {
210 std::string typeAtEncoding =
212 elements.add(CGM.GetAddrOfConstantCString(typeAtEncoding).getPointer());
213 }
214
215 // GC layout.
216 if (C.getLangOpts().ObjC) {
217 if (CGM.getLangOpts().getGC() != LangOptions::NonGC)
218 elements.add(CGM.getObjCRuntime().BuildGCBlockLayout(CGM, blockInfo));
219 else
220 elements.add(CGM.getObjCRuntime().BuildRCBlockLayout(CGM, blockInfo));
221 }
222 else
223 elements.addNullPointer(i8p);
224
225 unsigned AddrSpace = 0;
226 if (C.getLangOpts().OpenCL)
227 AddrSpace = C.getTargetAddressSpace(LangAS::opencl_constant);
228
229 llvm::GlobalValue::LinkageTypes linkage;
230 if (descName.empty()) {
231 linkage = llvm::GlobalValue::InternalLinkage;
232 descName = "__block_descriptor_tmp";
233 } else if (hasInternalHelper) {
234 // If either the copy helper or the dispose helper has internal linkage,
235 // the block descriptor must have internal linkage too.
236 linkage = llvm::GlobalValue::InternalLinkage;
237 } else {
238 linkage = llvm::GlobalValue::LinkOnceODRLinkage;
239 }
240
241 llvm::GlobalVariable *global =
242 elements.finishAndCreateGlobal(descName, CGM.getPointerAlign(),
243 /*constant*/ true, linkage, AddrSpace);
244
245 if (linkage == llvm::GlobalValue::LinkOnceODRLinkage) {
246 if (CGM.supportsCOMDAT())
247 global->setComdat(CGM.getModule().getOrInsertComdat(descName));
248 global->setVisibility(llvm::GlobalValue::HiddenVisibility);
249 global->setUnnamedAddr(llvm::GlobalValue::UnnamedAddr::Global);
250 }
251
252 return global;
253}
254
255/*
256 Purely notional variadic template describing the layout of a block.
257
258 template <class _ResultType, class... _ParamTypes, class... _CaptureTypes>
259 struct Block_literal {
260 /// Initialized to one of:
261 /// extern void *_NSConcreteStackBlock[];
262 /// extern void *_NSConcreteGlobalBlock[];
263 ///
264 /// In theory, we could start one off malloc'ed by setting
265 /// BLOCK_NEEDS_FREE, giving it a refcount of 1, and using
266 /// this isa:
267 /// extern void *_NSConcreteMallocBlock[];
268 struct objc_class *isa;
269
270 /// These are the flags (with corresponding bit number) that the
271 /// compiler is actually supposed to know about.
272 /// 23. BLOCK_IS_NOESCAPE - indicates that the block is non-escaping
273 /// 25. BLOCK_HAS_COPY_DISPOSE - indicates that the block
274 /// descriptor provides copy and dispose helper functions
275 /// 26. BLOCK_HAS_CXX_OBJ - indicates that there's a captured
276 /// object with a nontrivial destructor or copy constructor
277 /// 28. BLOCK_IS_GLOBAL - indicates that the block is allocated
278 /// as global memory
279 /// 29. BLOCK_USE_STRET - indicates that the block function
280 /// uses stret, which objc_msgSend needs to know about
281 /// 30. BLOCK_HAS_SIGNATURE - indicates that the block has an
282 /// @encoded signature string
283 /// And we're not supposed to manipulate these:
284 /// 24. BLOCK_NEEDS_FREE - indicates that the block has been moved
285 /// to malloc'ed memory
286 /// 27. BLOCK_IS_GC - indicates that the block has been moved to
287 /// to GC-allocated memory
288 /// Additionally, the bottom 16 bits are a reference count which
289 /// should be zero on the stack.
290 int flags;
291
292 /// Reserved; should be zero-initialized.
293 int reserved;
294
295 /// Function pointer generated from block literal.
296 _ResultType (*invoke)(Block_literal *, _ParamTypes...);
297
298 /// Block description metadata generated from block literal.
299 struct Block_descriptor *block_descriptor;
300
301 /// Captured values follow.
302 _CapturesTypes captures...;
303 };
304 */
305
306namespace {
307 /// A chunk of data that we actually have to capture in the block.
308 struct BlockLayoutChunk {
309 CharUnits Alignment;
311 const BlockDecl::Capture *Capture; // null for 'this'
312 llvm::Type *Type;
313 QualType FieldType;
314 BlockCaptureEntityKind CopyKind, DisposeKind;
315 BlockFieldFlags CopyFlags, DisposeFlags;
316
317 BlockLayoutChunk(CharUnits align, CharUnits size,
318 const BlockDecl::Capture *capture, llvm::Type *type,
319 QualType fieldType, BlockCaptureEntityKind CopyKind,
320 BlockFieldFlags CopyFlags,
321 BlockCaptureEntityKind DisposeKind,
322 BlockFieldFlags DisposeFlags)
323 : Alignment(align), Size(size), Capture(capture), Type(type),
324 FieldType(fieldType), CopyKind(CopyKind), DisposeKind(DisposeKind),
325 CopyFlags(CopyFlags), DisposeFlags(DisposeFlags) {}
326
327 /// Tell the block info that this chunk has the given field index.
328 void setIndex(CGBlockInfo &info, unsigned index, CharUnits offset) {
329 if (!Capture) {
330 info.CXXThisIndex = index;
331 info.CXXThisOffset = offset;
332 } else {
334 index, offset, FieldType, CopyKind, CopyFlags, DisposeKind,
335 DisposeFlags, Capture));
336 }
337 }
338
339 bool isTrivial() const {
340 return CopyKind == BlockCaptureEntityKind::None &&
341 DisposeKind == BlockCaptureEntityKind::None;
342 }
343 };
344
345 /// Order by 1) all __strong together 2) next, all block together 3) next,
346 /// all byref together 4) next, all __weak together. Preserve descending
347 /// alignment in all situations.
348 bool operator<(const BlockLayoutChunk &left, const BlockLayoutChunk &right) {
349 if (left.Alignment != right.Alignment)
350 return left.Alignment > right.Alignment;
351
352 auto getPrefOrder = [](const BlockLayoutChunk &chunk) {
353 switch (chunk.CopyKind) {
355 return 0;
357 switch (chunk.CopyFlags.getBitMask()) {
359 return 0;
361 return 1;
363 return 2;
364 default:
365 break;
366 }
367 break;
369 return 3;
370 default:
371 break;
372 }
373 return 4;
374 };
375
376 return getPrefOrder(left) < getPrefOrder(right);
377 }
378} // end anonymous namespace
379
380static std::pair<BlockCaptureEntityKind, BlockFieldFlags>
382 const LangOptions &LangOpts);
383
384static std::pair<BlockCaptureEntityKind, BlockFieldFlags>
386 const LangOptions &LangOpts);
387
388static void addBlockLayout(CharUnits align, CharUnits size,
389 const BlockDecl::Capture *capture, llvm::Type *type,
390 QualType fieldType,
392 CGBlockInfo &Info, CodeGenModule &CGM) {
393 if (!capture) {
394 // 'this' capture.
395 Layout.push_back(BlockLayoutChunk(
396 align, size, capture, type, fieldType, BlockCaptureEntityKind::None,
398 return;
399 }
400
401 const LangOptions &LangOpts = CGM.getLangOpts();
402 BlockCaptureEntityKind CopyKind, DisposeKind;
403 BlockFieldFlags CopyFlags, DisposeFlags;
404
405 std::tie(CopyKind, CopyFlags) =
406 computeCopyInfoForBlockCapture(*capture, fieldType, LangOpts);
407 std::tie(DisposeKind, DisposeFlags) =
408 computeDestroyInfoForBlockCapture(*capture, fieldType, LangOpts);
409 Layout.push_back(BlockLayoutChunk(align, size, capture, type, fieldType,
410 CopyKind, CopyFlags, DisposeKind,
411 DisposeFlags));
412
413 if (Info.NoEscape)
414 return;
415
416 if (!Layout.back().isTrivial())
417 Info.NeedsCopyDispose = true;
418}
419
420/// Determines if the given type is safe for constant capture in C++.
422 const RecordType *recordType =
423 type->getBaseElementTypeUnsafe()->getAs<RecordType>();
424
425 // Only records can be unsafe.
426 if (!recordType) return true;
427
428 const auto *record = cast<CXXRecordDecl>(recordType->getDecl());
429
430 // Maintain semantics for classes with non-trivial dtors or copy ctors.
431 if (!record->hasTrivialDestructor()) return false;
432 if (record->hasNonTrivialCopyConstructor()) return false;
433
434 // Otherwise, we just have to make sure there aren't any mutable
435 // fields that might have changed since initialization.
436 return !record->hasMutableFields();
437}
438
439/// It is illegal to modify a const object after initialization.
440/// Therefore, if a const object has a constant initializer, we don't
441/// actually need to keep storage for it in the block; we'll just
442/// rematerialize it at the start of the block function. This is
443/// acceptable because we make no promises about address stability of
444/// captured variables.
445static llvm::Constant *tryCaptureAsConstant(CodeGenModule &CGM,
446 CodeGenFunction *CGF,
447 const VarDecl *var) {
448 // Return if this is a function parameter. We shouldn't try to
449 // rematerialize default arguments of function parameters.
450 if (isa<ParmVarDecl>(var))
451 return nullptr;
452
453 QualType type = var->getType();
454
455 // We can only do this if the variable is const.
456 if (!type.isConstQualified()) return nullptr;
457
458 // Furthermore, in C++ we have to worry about mutable fields:
459 // C++ [dcl.type.cv]p4:
460 // Except that any class member declared mutable can be
461 // modified, any attempt to modify a const object during its
462 // lifetime results in undefined behavior.
463 if (CGM.getLangOpts().CPlusPlus && !isSafeForCXXConstantCapture(type))
464 return nullptr;
465
466 // If the variable doesn't have any initializer (shouldn't this be
467 // invalid?), it's not clear what we should do. Maybe capture as
468 // zero?
469 const Expr *init = var->getInit();
470 if (!init) return nullptr;
471
472 return ConstantEmitter(CGM, CGF).tryEmitAbstractForInitializer(*var);
473}
474
475/// Get the low bit of a nonzero character count. This is the
476/// alignment of the nth byte if the 0th byte is universally aligned.
478 return CharUnits::fromQuantity(v.getQuantity() & (~v.getQuantity() + 1));
479}
480
482 SmallVectorImpl<llvm::Type*> &elementTypes) {
483
484 assert(elementTypes.empty());
485 if (CGM.getLangOpts().OpenCL) {
486 // The header is basically 'struct { int; int; generic void *;
487 // custom_fields; }'. Assert that struct is packed.
488 auto GenPtrAlign = CharUnits::fromQuantity(
490 auto GenPtrSize = CharUnits::fromQuantity(
492 assert(CGM.getIntSize() <= GenPtrSize);
493 assert(CGM.getIntAlign() <= GenPtrAlign);
494 assert((2 * CGM.getIntSize()).isMultipleOf(GenPtrAlign));
495 elementTypes.push_back(CGM.IntTy); /* total size */
496 elementTypes.push_back(CGM.IntTy); /* align */
497 elementTypes.push_back(
498 CGM.getOpenCLRuntime()
499 .getGenericVoidPointerType()); /* invoke function */
500 unsigned Offset =
501 2 * CGM.getIntSize().getQuantity() + GenPtrSize.getQuantity();
502 unsigned BlockAlign = GenPtrAlign.getQuantity();
503 if (auto *Helper =
505 for (auto *I : Helper->getCustomFieldTypes()) /* custom fields */ {
506 // TargetOpenCLBlockHelp needs to make sure the struct is packed.
507 // If necessary, add padding fields to the custom fields.
508 unsigned Align = CGM.getDataLayout().getABITypeAlign(I).value();
509 if (BlockAlign < Align)
510 BlockAlign = Align;
511 assert(Offset % Align == 0);
512 Offset += CGM.getDataLayout().getTypeAllocSize(I);
513 elementTypes.push_back(I);
514 }
515 }
516 info.BlockAlign = CharUnits::fromQuantity(BlockAlign);
517 info.BlockSize = CharUnits::fromQuantity(Offset);
518 } else {
519 // The header is basically 'struct { void *; int; int; void *; void *; }'.
520 // Assert that the struct is packed.
521 assert(CGM.getIntSize() <= CGM.getPointerSize());
522 assert(CGM.getIntAlign() <= CGM.getPointerAlign());
523 assert((2 * CGM.getIntSize()).isMultipleOf(CGM.getPointerAlign()));
524 info.BlockAlign = CGM.getPointerAlign();
525 info.BlockSize = 3 * CGM.getPointerSize() + 2 * CGM.getIntSize();
526 elementTypes.push_back(CGM.VoidPtrTy);
527 elementTypes.push_back(CGM.IntTy);
528 elementTypes.push_back(CGM.IntTy);
529 elementTypes.push_back(CGM.VoidPtrTy);
530 elementTypes.push_back(CGM.getBlockDescriptorType());
531 }
532}
533
535 const BlockDecl::Capture &CI) {
536 const VarDecl *VD = CI.getVariable();
537
538 // If the variable is captured by an enclosing block or lambda expression,
539 // use the type of the capture field.
540 if (CGF.BlockInfo && CI.isNested())
541 return CGF.BlockInfo->getCapture(VD).fieldType();
542 if (auto *FD = CGF.LambdaCaptureFields.lookup(VD))
543 return FD->getType();
544 // If the captured variable is a non-escaping __block variable, the field
545 // type is the reference type. If the variable is a __block variable that
546 // already has a reference type, the field type is the variable's type.
547 return VD->isNonEscapingByref() ?
549}
550
551/// Compute the layout of the given block. Attempts to lay the block
552/// out with minimal space requirements.
554 CGBlockInfo &info) {
555 ASTContext &C = CGM.getContext();
556 const BlockDecl *block = info.getBlockDecl();
557
558 SmallVector<llvm::Type*, 8> elementTypes;
559 initializeForBlockHeader(CGM, info, elementTypes);
560 bool hasNonConstantCustomFields = false;
561 if (auto *OpenCLHelper =
563 hasNonConstantCustomFields =
564 !OpenCLHelper->areAllCustomFieldValuesConstant(info);
565 if (!block->hasCaptures() && !hasNonConstantCustomFields) {
566 info.StructureType =
567 llvm::StructType::get(CGM.getLLVMContext(), elementTypes, true);
568 info.CanBeGlobal = true;
569 return;
570 }
571 else if (C.getLangOpts().ObjC &&
572 CGM.getLangOpts().getGC() == LangOptions::NonGC)
573 info.HasCapturedVariableLayout = true;
574
575 if (block->doesNotEscape())
576 info.NoEscape = true;
577
578 // Collect the layout chunks.
580 layout.reserve(block->capturesCXXThis() +
581 (block->capture_end() - block->capture_begin()));
582
583 CharUnits maxFieldAlign;
584
585 // First, 'this'.
586 if (block->capturesCXXThis()) {
587 assert(CGF && isa_and_nonnull<CXXMethodDecl>(CGF->CurFuncDecl) &&
588 "Can't capture 'this' outside a method");
589 QualType thisType = cast<CXXMethodDecl>(CGF->CurFuncDecl)->getThisType();
590
591 // Theoretically, this could be in a different address space, so
592 // don't assume standard pointer size/align.
593 llvm::Type *llvmType = CGM.getTypes().ConvertType(thisType);
594 auto TInfo = CGM.getContext().getTypeInfoInChars(thisType);
595 maxFieldAlign = std::max(maxFieldAlign, TInfo.Align);
596
597 addBlockLayout(TInfo.Align, TInfo.Width, nullptr, llvmType, thisType,
598 layout, info, CGM);
599 }
600
601 // Next, all the block captures.
602 for (const auto &CI : block->captures()) {
603 const VarDecl *variable = CI.getVariable();
604
605 if (CI.isEscapingByref()) {
606 // Just use void* instead of a pointer to the byref type.
607 CharUnits align = CGM.getPointerAlign();
608 maxFieldAlign = std::max(maxFieldAlign, align);
609
610 // Since a __block variable cannot be captured by lambdas, its type and
611 // the capture field type should always match.
612 assert(CGF && getCaptureFieldType(*CGF, CI) == variable->getType() &&
613 "capture type differs from the variable type");
614 addBlockLayout(align, CGM.getPointerSize(), &CI, CGM.VoidPtrTy,
615 variable->getType(), layout, info, CGM);
616 continue;
617 }
618
619 // Otherwise, build a layout chunk with the size and alignment of
620 // the declaration.
621 if (llvm::Constant *constant = tryCaptureAsConstant(CGM, CGF, variable)) {
622 info.SortedCaptures.push_back(
624 continue;
625 }
626
627 QualType VT = getCaptureFieldType(*CGF, CI);
628
629 if (CGM.getLangOpts().CPlusPlus)
630 if (const CXXRecordDecl *record = VT->getAsCXXRecordDecl())
631 if (CI.hasCopyExpr() || !record->hasTrivialDestructor()) {
632 info.HasCXXObject = true;
633 if (!record->isExternallyVisible())
634 info.CapturesNonExternalType = true;
635 }
636
637 CharUnits size = C.getTypeSizeInChars(VT);
638 CharUnits align = C.getDeclAlign(variable);
639
640 maxFieldAlign = std::max(maxFieldAlign, align);
641
642 llvm::Type *llvmType =
643 CGM.getTypes().ConvertTypeForMem(VT);
644
645 addBlockLayout(align, size, &CI, llvmType, VT, layout, info, CGM);
646 }
647
648 // If that was everything, we're done here.
649 if (layout.empty()) {
650 info.StructureType =
651 llvm::StructType::get(CGM.getLLVMContext(), elementTypes, true);
652 info.CanBeGlobal = true;
653 info.buildCaptureMap();
654 return;
655 }
656
657 // Sort the layout by alignment. We have to use a stable sort here
658 // to get reproducible results. There should probably be an
659 // llvm::array_pod_stable_sort.
660 llvm::stable_sort(layout);
661
662 // Needed for blocks layout info.
665
666 CharUnits &blockSize = info.BlockSize;
667 info.BlockAlign = std::max(maxFieldAlign, info.BlockAlign);
668
669 // Assuming that the first byte in the header is maximally aligned,
670 // get the alignment of the first byte following the header.
671 CharUnits endAlign = getLowBit(blockSize);
672
673 // If the end of the header isn't satisfactorily aligned for the
674 // maximum thing, look for things that are okay with the header-end
675 // alignment, and keep appending them until we get something that's
676 // aligned right. This algorithm is only guaranteed optimal if
677 // that condition is satisfied at some point; otherwise we can get
678 // things like:
679 // header // next byte has alignment 4
680 // something_with_size_5; // next byte has alignment 1
681 // something_with_alignment_8;
682 // which has 7 bytes of padding, as opposed to the naive solution
683 // which might have less (?).
684 if (endAlign < maxFieldAlign) {
686 li = layout.begin() + 1, le = layout.end();
687
688 // Look for something that the header end is already
689 // satisfactorily aligned for.
690 for (; li != le && endAlign < li->Alignment; ++li)
691 ;
692
693 // If we found something that's naturally aligned for the end of
694 // the header, keep adding things...
695 if (li != le) {
697 for (; li != le; ++li) {
698 assert(endAlign >= li->Alignment);
699
700 li->setIndex(info, elementTypes.size(), blockSize);
701 elementTypes.push_back(li->Type);
702 blockSize += li->Size;
703 endAlign = getLowBit(blockSize);
704
705 // ...until we get to the alignment of the maximum field.
706 if (endAlign >= maxFieldAlign) {
707 ++li;
708 break;
709 }
710 }
711 // Don't re-append everything we just appended.
712 layout.erase(first, li);
713 }
714 }
715
716 assert(endAlign == getLowBit(blockSize));
717
718 // At this point, we just have to add padding if the end align still
719 // isn't aligned right.
720 if (endAlign < maxFieldAlign) {
721 CharUnits newBlockSize = blockSize.alignTo(maxFieldAlign);
722 CharUnits padding = newBlockSize - blockSize;
723
724 // If we haven't yet added any fields, remember that there was an
725 // initial gap; this need to go into the block layout bit map.
726 if (blockSize == info.BlockHeaderForcedGapOffset) {
727 info.BlockHeaderForcedGapSize = padding;
728 }
729
730 elementTypes.push_back(llvm::ArrayType::get(CGM.Int8Ty,
731 padding.getQuantity()));
732 blockSize = newBlockSize;
733 endAlign = getLowBit(blockSize); // might be > maxFieldAlign
734 }
735
736 assert(endAlign >= maxFieldAlign);
737 assert(endAlign == getLowBit(blockSize));
738 // Slam everything else on now. This works because they have
739 // strictly decreasing alignment and we expect that size is always a
740 // multiple of alignment.
742 li = layout.begin(), le = layout.end(); li != le; ++li) {
743 if (endAlign < li->Alignment) {
744 // size may not be multiple of alignment. This can only happen with
745 // an over-aligned variable. We will be adding a padding field to
746 // make the size be multiple of alignment.
747 CharUnits padding = li->Alignment - endAlign;
748 elementTypes.push_back(llvm::ArrayType::get(CGM.Int8Ty,
749 padding.getQuantity()));
750 blockSize += padding;
751 endAlign = getLowBit(blockSize);
752 }
753 assert(endAlign >= li->Alignment);
754 li->setIndex(info, elementTypes.size(), blockSize);
755 elementTypes.push_back(li->Type);
756 blockSize += li->Size;
757 endAlign = getLowBit(blockSize);
758 }
759
760 info.buildCaptureMap();
761 info.StructureType =
762 llvm::StructType::get(CGM.getLLVMContext(), elementTypes, true);
763}
764
765/// Emit a block literal expression in the current function.
767 // If the block has no captures, we won't have a pre-computed
768 // layout for it.
769 if (!blockExpr->getBlockDecl()->hasCaptures())
770 // The block literal is emitted as a global variable, and the block invoke
771 // function has to be extracted from its initializer.
772 if (llvm::Constant *Block = CGM.getAddrOfGlobalBlockIfEmitted(blockExpr))
773 return Block;
774
775 CGBlockInfo blockInfo(blockExpr->getBlockDecl(), CurFn->getName());
776 computeBlockInfo(CGM, this, blockInfo);
777 blockInfo.BlockExpression = blockExpr;
778 if (!blockInfo.CanBeGlobal)
779 blockInfo.LocalAddress = CreateTempAlloca(blockInfo.StructureType,
780 blockInfo.BlockAlign, "block");
781 return EmitBlockLiteral(blockInfo);
782}
783
784llvm::Value *CodeGenFunction::EmitBlockLiteral(const CGBlockInfo &blockInfo) {
785 bool IsOpenCL = CGM.getContext().getLangOpts().OpenCL;
786 auto GenVoidPtrTy =
788 LangAS GenVoidPtrAddr = IsOpenCL ? LangAS::opencl_generic : LangAS::Default;
789 auto GenVoidPtrSize = CharUnits::fromQuantity(
790 CGM.getTarget().getPointerWidth(GenVoidPtrAddr) / 8);
791 // Using the computed layout, generate the actual block function.
792 bool isLambdaConv = blockInfo.getBlockDecl()->isConversionFromLambda();
793 CodeGenFunction BlockCGF{CGM, true};
794 BlockCGF.SanOpts = SanOpts;
795 auto *InvokeFn = BlockCGF.GenerateBlockFunction(
796 CurGD, blockInfo, LocalDeclMap, isLambdaConv, blockInfo.CanBeGlobal);
797 auto *blockFn = llvm::ConstantExpr::getPointerCast(InvokeFn, GenVoidPtrTy);
798
799 // If there is nothing to capture, we can emit this as a global block.
800 if (blockInfo.CanBeGlobal)
802
803 // Otherwise, we have to emit this as a local block.
804
805 RawAddress blockAddr = blockInfo.LocalAddress;
806 assert(blockAddr.isValid() && "block has no address!");
807
808 llvm::Constant *isa;
809 llvm::Constant *descriptor;
810 BlockFlags flags;
811 if (!IsOpenCL) {
812 // If the block is non-escaping, set field 'isa 'to NSConcreteGlobalBlock
813 // and set the BLOCK_IS_GLOBAL bit of field 'flags'. Copying a non-escaping
814 // block just returns the original block and releasing it is a no-op.
815 llvm::Constant *blockISA = blockInfo.NoEscape
818 isa = blockISA;
819
820 // Build the block descriptor.
821 descriptor = buildBlockDescriptor(CGM, blockInfo);
822
823 // Compute the initial on-stack block flags.
824 if (!CGM.getCodeGenOpts().DisableBlockSignatureString)
825 flags = BLOCK_HAS_SIGNATURE;
826 if (blockInfo.HasCapturedVariableLayout)
828 if (blockInfo.NeedsCopyDispose)
829 flags |= BLOCK_HAS_COPY_DISPOSE;
830 if (blockInfo.HasCXXObject)
831 flags |= BLOCK_HAS_CXX_OBJ;
832 if (blockInfo.UsesStret)
833 flags |= BLOCK_USE_STRET;
834 if (blockInfo.NoEscape)
836 }
837
838 auto projectField = [&](unsigned index, const Twine &name) -> Address {
839 return Builder.CreateStructGEP(blockAddr, index, name);
840 };
841 auto storeField = [&](llvm::Value *value, unsigned index, const Twine &name) {
842 Builder.CreateStore(value, projectField(index, name));
843 };
844
845 // Initialize the block header.
846 {
847 // We assume all the header fields are densely packed.
848 unsigned index = 0;
849 CharUnits offset;
850 auto addHeaderField = [&](llvm::Value *value, CharUnits size,
851 const Twine &name) {
852 storeField(value, index, name);
853 offset += size;
854 index++;
855 };
856
857 if (!IsOpenCL) {
858 addHeaderField(isa, getPointerSize(), "block.isa");
859 addHeaderField(llvm::ConstantInt::get(IntTy, flags.getBitMask()),
860 getIntSize(), "block.flags");
861 addHeaderField(llvm::ConstantInt::get(IntTy, 0), getIntSize(),
862 "block.reserved");
863 } else {
864 addHeaderField(
865 llvm::ConstantInt::get(IntTy, blockInfo.BlockSize.getQuantity()),
866 getIntSize(), "block.size");
867 addHeaderField(
868 llvm::ConstantInt::get(IntTy, blockInfo.BlockAlign.getQuantity()),
869 getIntSize(), "block.align");
870 }
871 addHeaderField(blockFn, GenVoidPtrSize, "block.invoke");
872 if (!IsOpenCL)
873 addHeaderField(descriptor, getPointerSize(), "block.descriptor");
874 else if (auto *Helper =
876 for (auto I : Helper->getCustomFieldValues(*this, blockInfo)) {
877 addHeaderField(
878 I.first,
880 CGM.getDataLayout().getTypeAllocSize(I.first->getType())),
881 I.second);
882 }
883 }
884 }
885
886 // Finally, capture all the values into the block.
887 const BlockDecl *blockDecl = blockInfo.getBlockDecl();
888
889 // First, 'this'.
890 if (blockDecl->capturesCXXThis()) {
891 Address addr =
892 projectField(blockInfo.CXXThisIndex, "block.captured-this.addr");
894 }
895
896 // Next, captured variables.
897 for (const auto &CI : blockDecl->captures()) {
898 const VarDecl *variable = CI.getVariable();
899 const CGBlockInfo::Capture &capture = blockInfo.getCapture(variable);
900
901 // Ignore constant captures.
902 if (capture.isConstant()) continue;
903
904 QualType type = capture.fieldType();
905
906 // This will be a [[type]]*, except that a byref entry will just be
907 // an i8**.
908 Address blockField = projectField(capture.getIndex(), "block.captured");
909
910 // Compute the address of the thing we're going to move into the
911 // block literal.
913
914 if (blockDecl->isConversionFromLambda()) {
915 // The lambda capture in a lambda's conversion-to-block-pointer is
916 // special; we'll simply emit it directly.
917 src = Address::invalid();
918 } else if (CI.isEscapingByref()) {
919 if (BlockInfo && CI.isNested()) {
920 // We need to use the capture from the enclosing block.
921 const CGBlockInfo::Capture &enclosingCapture =
922 BlockInfo->getCapture(variable);
923
924 // This is a [[type]]*, except that a byref entry will just be an i8**.
926 enclosingCapture.getIndex(),
927 "block.capture.addr");
928 } else {
929 auto I = LocalDeclMap.find(variable);
930 assert(I != LocalDeclMap.end());
931 src = I->second;
932 }
933 } else {
934 DeclRefExpr declRef(getContext(), const_cast<VarDecl *>(variable),
935 /*RefersToEnclosingVariableOrCapture*/ CI.isNested(),
936 type.getNonReferenceType(), VK_LValue,
938 src = EmitDeclRefLValue(&declRef).getAddress();
939 };
940
941 // For byrefs, we just write the pointer to the byref struct into
942 // the block field. There's no need to chase the forwarding
943 // pointer at this point, since we're building something that will
944 // live a shorter life than the stack byref anyway.
945 if (CI.isEscapingByref()) {
946 // Get a void* that points to the byref struct.
947 llvm::Value *byrefPointer;
948 if (CI.isNested())
949 byrefPointer = Builder.CreateLoad(src, "byref.capture");
950 else
951 byrefPointer = src.emitRawPointer(*this);
952
953 // Write that void* into the capture field.
954 Builder.CreateStore(byrefPointer, blockField);
955
956 // If we have a copy constructor, evaluate that into the block field.
957 } else if (const Expr *copyExpr = CI.getCopyExpr()) {
958 if (blockDecl->isConversionFromLambda()) {
959 // If we have a lambda conversion, emit the expression
960 // directly into the block instead.
961 AggValueSlot Slot =
962 AggValueSlot::forAddr(blockField, Qualifiers(),
967 EmitAggExpr(copyExpr, Slot);
968 } else {
969 EmitSynthesizedCXXCopyCtor(blockField, src, copyExpr);
970 }
971
972 // If it's a reference variable, copy the reference into the block field.
973 } else if (type->getAs<ReferenceType>()) {
974 Builder.CreateStore(src.emitRawPointer(*this), blockField);
975
976 // If type is const-qualified, copy the value into the block field.
977 } else if (type.isConstQualified() &&
978 type.getObjCLifetime() == Qualifiers::OCL_Strong &&
979 CGM.getCodeGenOpts().OptimizationLevel != 0) {
980 llvm::Value *value = Builder.CreateLoad(src, "captured");
981 Builder.CreateStore(value, blockField);
982
983 // If this is an ARC __strong block-pointer variable, don't do a
984 // block copy.
985 //
986 // TODO: this can be generalized into the normal initialization logic:
987 // we should never need to do a block-copy when initializing a local
988 // variable, because the local variable's lifetime should be strictly
989 // contained within the stack block's.
990 } else if (type.getObjCLifetime() == Qualifiers::OCL_Strong &&
991 type->isBlockPointerType()) {
992 // Load the block and do a simple retain.
993 llvm::Value *value = Builder.CreateLoad(src, "block.captured_block");
994 value = EmitARCRetainNonBlock(value);
995
996 // Do a primitive store to the block field.
997 Builder.CreateStore(value, blockField);
998
999 // Otherwise, fake up a POD copy into the block field.
1000 } else {
1001 // Fake up a new variable so that EmitScalarInit doesn't think
1002 // we're referring to the variable in its own initializer.
1003 ImplicitParamDecl BlockFieldPseudoVar(getContext(), type,
1005
1006 // We use one of these or the other depending on whether the
1007 // reference is nested.
1008 DeclRefExpr declRef(getContext(), const_cast<VarDecl *>(variable),
1009 /*RefersToEnclosingVariableOrCapture*/ CI.isNested(),
1011
1012 ImplicitCastExpr l2r(ImplicitCastExpr::OnStack, type, CK_LValueToRValue,
1013 &declRef, VK_PRValue, FPOptionsOverride());
1014 // FIXME: Pass a specific location for the expr init so that the store is
1015 // attributed to a reasonable location - otherwise it may be attributed to
1016 // locations of subexpressions in the initialization.
1017 EmitExprAsInit(&l2r, &BlockFieldPseudoVar,
1019 /*captured by init*/ false);
1020 }
1021
1022 // Push a cleanup for the capture if necessary.
1023 if (!blockInfo.NoEscape && !blockInfo.NeedsCopyDispose)
1024 continue;
1025
1026 // Ignore __block captures; there's nothing special in the on-stack block
1027 // that we need to do for them.
1028 if (CI.isByRef())
1029 continue;
1030
1031 // Ignore objects that aren't destructed.
1032 QualType::DestructionKind dtorKind = type.isDestructedType();
1033 if (dtorKind == QualType::DK_none)
1034 continue;
1035
1036 CodeGenFunction::Destroyer *destroyer;
1037
1038 // Block captures count as local values and have imprecise semantics.
1039 // They also can't be arrays, so need to worry about that.
1040 //
1041 // For const-qualified captures, emit clang.arc.use to ensure the captured
1042 // object doesn't get released while we are still depending on its validity
1043 // within the block.
1044 if (type.isConstQualified() &&
1045 type.getObjCLifetime() == Qualifiers::OCL_Strong &&
1046 CGM.getCodeGenOpts().OptimizationLevel != 0) {
1047 assert(CGM.getLangOpts().ObjCAutoRefCount &&
1048 "expected ObjC ARC to be enabled");
1049 destroyer = emitARCIntrinsicUse;
1050 } else if (dtorKind == QualType::DK_objc_strong_lifetime) {
1051 destroyer = destroyARCStrongImprecise;
1052 } else {
1053 destroyer = getDestroyer(dtorKind);
1054 }
1055
1056 CleanupKind cleanupKind = NormalCleanup;
1057 bool useArrayEHCleanup = needsEHCleanup(dtorKind);
1058 if (useArrayEHCleanup)
1059 cleanupKind = NormalAndEHCleanup;
1060
1061 // Extend the lifetime of the capture to the end of the scope enclosing the
1062 // block expression except when the block decl is in the list of RetExpr's
1063 // cleanup objects, in which case its lifetime ends after the full
1064 // expression.
1065 auto IsBlockDeclInRetExpr = [&]() {
1066 auto *EWC = llvm::dyn_cast_or_null<ExprWithCleanups>(RetExpr);
1067 if (EWC)
1068 for (auto &C : EWC->getObjects())
1069 if (auto *BD = C.dyn_cast<BlockDecl *>())
1070 if (BD == blockDecl)
1071 return true;
1072 return false;
1073 };
1074
1075 if (IsBlockDeclInRetExpr())
1076 pushDestroy(cleanupKind, blockField, type, destroyer, useArrayEHCleanup);
1077 else
1078 pushLifetimeExtendedDestroy(cleanupKind, blockField, type, destroyer,
1079 useArrayEHCleanup);
1080 }
1081
1082 // Cast to the converted block-pointer type, which happens (somewhat
1083 // unfortunately) to be a pointer to function type.
1084 llvm::Value *result = Builder.CreatePointerCast(
1085 blockAddr.getPointer(), ConvertType(blockInfo.getBlockExpr()->getType()));
1086
1087 if (IsOpenCL) {
1089 result, blockInfo.StructureType);
1090 }
1091
1092 return result;
1093}
1094
1095
1097 if (BlockDescriptorType)
1098 return BlockDescriptorType;
1099
1100 llvm::Type *UnsignedLongTy =
1101 getTypes().ConvertType(getContext().UnsignedLongTy);
1102
1103 // struct __block_descriptor {
1104 // unsigned long reserved;
1105 // unsigned long block_size;
1106 //
1107 // // later, the following will be added
1108 //
1109 // struct {
1110 // void (*copyHelper)();
1111 // void (*copyHelper)();
1112 // } helpers; // !!! optional
1113 //
1114 // const char *signature; // the block signature
1115 // const char *layout; // reserved
1116 // };
1117 BlockDescriptorType = llvm::StructType::create(
1118 "struct.__block_descriptor", UnsignedLongTy, UnsignedLongTy);
1119
1120 // Now form a pointer to that.
1121 unsigned AddrSpace = 0;
1122 if (getLangOpts().OpenCL)
1124 BlockDescriptorType = llvm::PointerType::get(BlockDescriptorType, AddrSpace);
1125 return BlockDescriptorType;
1126}
1127
1129 if (GenericBlockLiteralType)
1130 return GenericBlockLiteralType;
1131
1132 llvm::Type *BlockDescPtrTy = getBlockDescriptorType();
1133
1134 if (getLangOpts().OpenCL) {
1135 // struct __opencl_block_literal_generic {
1136 // int __size;
1137 // int __align;
1138 // __generic void *__invoke;
1139 // /* custom fields */
1140 // };
1141 SmallVector<llvm::Type *, 8> StructFields(
1143 if (auto *Helper = getTargetCodeGenInfo().getTargetOpenCLBlockHelper()) {
1144 llvm::append_range(StructFields, Helper->getCustomFieldTypes());
1145 }
1146 GenericBlockLiteralType = llvm::StructType::create(
1147 StructFields, "struct.__opencl_block_literal_generic");
1148 } else {
1149 // struct __block_literal_generic {
1150 // void *__isa;
1151 // int __flags;
1152 // int __reserved;
1153 // void (*__invoke)(void *);
1154 // struct __block_descriptor *__descriptor;
1155 // };
1156 GenericBlockLiteralType =
1157 llvm::StructType::create("struct.__block_literal_generic", VoidPtrTy,
1158 IntTy, IntTy, VoidPtrTy, BlockDescPtrTy);
1159 }
1160
1161 return GenericBlockLiteralType;
1162}
1163
1165 ReturnValueSlot ReturnValue,
1166 llvm::CallBase **CallOrInvoke) {
1167 const auto *BPT = E->getCallee()->getType()->castAs<BlockPointerType>();
1168 llvm::Value *BlockPtr = EmitScalarExpr(E->getCallee());
1169 llvm::Type *GenBlockTy = CGM.getGenericBlockLiteralType();
1170 llvm::Value *Func = nullptr;
1171 QualType FnType = BPT->getPointeeType();
1172 ASTContext &Ctx = getContext();
1173 CallArgList Args;
1174
1175 if (getLangOpts().OpenCL) {
1176 // For OpenCL, BlockPtr is already casted to generic block literal.
1177
1178 // First argument of a block call is a generic block literal casted to
1179 // generic void pointer, i.e. i8 addrspace(4)*
1180 llvm::Type *GenericVoidPtrTy =
1182 llvm::Value *BlockDescriptor = Builder.CreatePointerCast(
1183 BlockPtr, GenericVoidPtrTy);
1184 QualType VoidPtrQualTy = Ctx.getPointerType(
1186 Args.add(RValue::get(BlockDescriptor), VoidPtrQualTy);
1187 // And the rest of the arguments.
1188 EmitCallArgs(Args, FnType->getAs<FunctionProtoType>(), E->arguments());
1189
1190 // We *can* call the block directly unless it is a function argument.
1191 if (!isa<ParmVarDecl>(E->getCalleeDecl()))
1192 Func = CGM.getOpenCLRuntime().getInvokeFunction(E->getCallee());
1193 else {
1194 llvm::Value *FuncPtr = Builder.CreateStructGEP(GenBlockTy, BlockPtr, 2);
1195 Func = Builder.CreateAlignedLoad(GenericVoidPtrTy, FuncPtr,
1196 getPointerAlign());
1197 }
1198 } else {
1199 // Bitcast the block literal to a generic block literal.
1200 BlockPtr =
1201 Builder.CreatePointerCast(BlockPtr, UnqualPtrTy, "block.literal");
1202 // Get pointer to the block invoke function
1203 llvm::Value *FuncPtr = Builder.CreateStructGEP(GenBlockTy, BlockPtr, 3);
1204
1205 // First argument is a block literal casted to a void pointer
1206 BlockPtr = Builder.CreatePointerCast(BlockPtr, VoidPtrTy);
1207 Args.add(RValue::get(BlockPtr), Ctx.VoidPtrTy);
1208 // And the rest of the arguments.
1209 EmitCallArgs(Args, FnType->getAs<FunctionProtoType>(), E->arguments());
1210
1211 // Load the function.
1213 }
1214
1215 const FunctionType *FuncTy = FnType->castAs<FunctionType>();
1216 const CGFunctionInfo &FnInfo =
1217 CGM.getTypes().arrangeBlockFunctionCall(Args, FuncTy);
1218
1219 // Prepare the callee.
1221
1222 // And call the block.
1223 return EmitCall(FnInfo, Callee, ReturnValue, Args, CallOrInvoke);
1224}
1225
1227 assert(BlockInfo && "evaluating block ref without block information?");
1228 const CGBlockInfo::Capture &capture = BlockInfo->getCapture(variable);
1229
1230 // Handle constant captures.
1231 if (capture.isConstant()) return LocalDeclMap.find(variable)->second;
1232
1234 "block.capture.addr");
1235
1236 if (variable->isEscapingByref()) {
1237 // addr should be a void** right now. Load, then cast the result
1238 // to byref*.
1239
1240 auto &byrefInfo = getBlockByrefInfo(variable);
1241 addr = Address(Builder.CreateLoad(addr), byrefInfo.Type,
1242 byrefInfo.ByrefAlignment);
1243
1244 addr = emitBlockByrefAddress(addr, byrefInfo, /*follow*/ true,
1245 variable->getName());
1246 }
1247
1248 assert((!variable->isNonEscapingByref() ||
1249 capture.fieldType()->isReferenceType()) &&
1250 "the capture field of a non-escaping variable should have a "
1251 "reference type");
1252 if (capture.fieldType()->isReferenceType())
1253 addr = EmitLoadOfReference(MakeAddrLValue(addr, capture.fieldType()));
1254
1255 return addr;
1256}
1257
1259 llvm::Constant *Addr) {
1260 bool Ok = EmittedGlobalBlocks.insert(std::make_pair(BE, Addr)).second;
1261 (void)Ok;
1262 assert(Ok && "Trying to replace an already-existing global block!");
1263}
1264
1265llvm::Constant *
1267 StringRef Name) {
1268 if (llvm::Constant *Block = getAddrOfGlobalBlockIfEmitted(BE))
1269 return Block;
1270
1271 CGBlockInfo blockInfo(BE->getBlockDecl(), Name);
1272 blockInfo.BlockExpression = BE;
1273
1274 // Compute information about the layout, etc., of this block.
1275 computeBlockInfo(*this, nullptr, blockInfo);
1276
1277 // Using that metadata, generate the actual block function.
1278 {
1279 CodeGenFunction::DeclMapTy LocalDeclMap;
1281 GlobalDecl(), blockInfo, LocalDeclMap,
1282 /*IsLambdaConversionToBlock*/ false, /*BuildGlobalBlock*/ true);
1283 }
1284
1286}
1287
1288static llvm::Constant *buildGlobalBlock(CodeGenModule &CGM,
1289 const CGBlockInfo &blockInfo,
1290 llvm::Constant *blockFn) {
1291 assert(blockInfo.CanBeGlobal);
1292 // Callers should detect this case on their own: calling this function
1293 // generally requires computing layout information, which is a waste of time
1294 // if we've already emitted this block.
1295 assert(!CGM.getAddrOfGlobalBlockIfEmitted(blockInfo.BlockExpression) &&
1296 "Refusing to re-emit a global block.");
1297
1298 // Generate the constants for the block literal initializer.
1299 ConstantInitBuilder builder(CGM);
1300 auto fields = builder.beginStruct();
1301
1302 bool IsOpenCL = CGM.getLangOpts().OpenCL;
1303 bool IsWindows = CGM.getTarget().getTriple().isOSWindows();
1304 if (!IsOpenCL) {
1305 // isa
1306 if (IsWindows)
1307 fields.addNullPointer(CGM.Int8PtrPtrTy);
1308 else
1309 fields.add(CGM.getNSConcreteGlobalBlock());
1310
1311 // __flags
1313 if (!CGM.getCodeGenOpts().DisableBlockSignatureString)
1314 flags |= BLOCK_HAS_SIGNATURE;
1315 if (blockInfo.UsesStret)
1316 flags |= BLOCK_USE_STRET;
1317
1318 fields.addInt(CGM.IntTy, flags.getBitMask());
1319
1320 // Reserved
1321 fields.addInt(CGM.IntTy, 0);
1322 } else {
1323 fields.addInt(CGM.IntTy, blockInfo.BlockSize.getQuantity());
1324 fields.addInt(CGM.IntTy, blockInfo.BlockAlign.getQuantity());
1325 }
1326
1327 // Function
1328 fields.add(blockFn);
1329
1330 if (!IsOpenCL) {
1331 // Descriptor
1332 fields.add(buildBlockDescriptor(CGM, blockInfo));
1333 } else if (auto *Helper =
1335 for (auto *I : Helper->getCustomFieldValues(CGM, blockInfo)) {
1336 fields.add(I);
1337 }
1338 }
1339
1340 unsigned AddrSpace = 0;
1341 if (CGM.getContext().getLangOpts().OpenCL)
1343
1344 llvm::GlobalVariable *literal = fields.finishAndCreateGlobal(
1345 "__block_literal_global", blockInfo.BlockAlign,
1346 /*constant*/ !IsWindows, llvm::GlobalVariable::InternalLinkage, AddrSpace);
1347
1348 literal->addAttribute("objc_arc_inert");
1349
1350 // Windows does not allow globals to be initialised to point to globals in
1351 // different DLLs. Any such variables must run code to initialise them.
1352 if (IsWindows) {
1353 auto *Init = llvm::Function::Create(llvm::FunctionType::get(CGM.VoidTy,
1354 {}), llvm::GlobalValue::InternalLinkage, ".block_isa_init",
1355 &CGM.getModule());
1356 llvm::IRBuilder<> b(llvm::BasicBlock::Create(CGM.getLLVMContext(), "entry",
1357 Init));
1358 b.CreateAlignedStore(CGM.getNSConcreteGlobalBlock(),
1359 b.CreateStructGEP(literal->getValueType(), literal, 0),
1360 CGM.getPointerAlign().getAsAlign());
1361 b.CreateRetVoid();
1362 // We can't use the normal LLVM global initialisation array, because we
1363 // need to specify that this runs early in library initialisation.
1364 auto *InitVar = new llvm::GlobalVariable(CGM.getModule(), Init->getType(),
1365 /*isConstant*/true, llvm::GlobalValue::InternalLinkage,
1366 Init, ".block_isa_init_ptr");
1367 InitVar->setSection(".CRT$XCLa");
1368 CGM.addUsedGlobal(InitVar);
1369 }
1370
1371 // Return a constant of the appropriately-casted type.
1372 llvm::Type *RequiredType =
1373 CGM.getTypes().ConvertType(blockInfo.getBlockExpr()->getType());
1374 llvm::Constant *Result =
1375 llvm::ConstantExpr::getPointerCast(literal, RequiredType);
1377 if (CGM.getContext().getLangOpts().OpenCL)
1379 blockInfo.BlockExpression,
1380 cast<llvm::Function>(blockFn->stripPointerCasts()), Result,
1381 literal->getValueType());
1382 return Result;
1383}
1384
1386 unsigned argNum,
1387 llvm::Value *arg) {
1388 assert(BlockInfo && "not emitting prologue of block invocation function?!");
1389
1390 // Allocate a stack slot like for any local variable to guarantee optimal
1391 // debug info at -O0. The mem2reg pass will eliminate it when optimizing.
1392 RawAddress alloc = CreateMemTemp(D->getType(), D->getName() + ".addr");
1393 Builder.CreateStore(arg, alloc);
1394 if (CGDebugInfo *DI = getDebugInfo()) {
1396 DI->setLocation(D->getLocation());
1397 DI->EmitDeclareOfBlockLiteralArgVariable(
1398 *BlockInfo, D->getName(), argNum,
1399 cast<llvm::AllocaInst>(alloc.getPointer()->stripPointerCasts()),
1400 Builder);
1401 }
1402 }
1403
1405 ApplyDebugLocation Scope(*this, StartLoc);
1406
1407 // Instead of messing around with LocalDeclMap, just set the value
1408 // directly as BlockPointer.
1409 BlockPointer = Builder.CreatePointerCast(
1410 arg,
1411 llvm::PointerType::get(
1414 ? getContext().getTargetAddressSpace(LangAS::opencl_generic)
1415 : 0),
1416 "block");
1417}
1418
1420 assert(BlockInfo && "not in a block invocation function!");
1421 assert(BlockPointer && "no block pointer set!");
1422 return Address(BlockPointer,