blob: 5800ce770c3e6267ea5b48a65b48c63704874415 [file] [log] [blame]
Shih-wei Liaof8fd82b2010-02-10 11:10:31 -08001//===--- CGExprConstant.cpp - Emit LLVM Code from Constant Expressions ----===//
2//
3// The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10// This contains code to emit Constant Expr nodes as LLVM code.
11//
12//===----------------------------------------------------------------------===//
13
14#include "CodeGenFunction.h"
15#include "CodeGenModule.h"
16#include "CGObjCRuntime.h"
17#include "clang/AST/APValue.h"
18#include "clang/AST/ASTContext.h"
19#include "clang/AST/RecordLayout.h"
20#include "clang/AST/StmtVisitor.h"
21#include "clang/Basic/Builtins.h"
22#include "llvm/Constants.h"
23#include "llvm/Function.h"
24#include "llvm/GlobalVariable.h"
25#include "llvm/Target/TargetData.h"
26using namespace clang;
27using namespace CodeGen;
28
29namespace {
30class ConstStructBuilder {
31 CodeGenModule &CGM;
32 CodeGenFunction *CGF;
33
34 bool Packed;
35
36 unsigned NextFieldOffsetInBytes;
37
38 unsigned LLVMStructAlignment;
39
40 std::vector<llvm::Constant *> Elements;
41
42 ConstStructBuilder(CodeGenModule &CGM, CodeGenFunction *CGF)
43 : CGM(CGM), CGF(CGF), Packed(false), NextFieldOffsetInBytes(0),
44 LLVMStructAlignment(1) { }
45
46 bool AppendField(const FieldDecl *Field, uint64_t FieldOffset,
47 const Expr *InitExpr) {
48 uint64_t FieldOffsetInBytes = FieldOffset / 8;
49
50 assert(NextFieldOffsetInBytes <= FieldOffsetInBytes
51 && "Field offset mismatch!");
52
53 // Emit the field.
54 llvm::Constant *C = CGM.EmitConstantExpr(InitExpr, Field->getType(), CGF);
55 if (!C)
56 return false;
57
58 unsigned FieldAlignment = getAlignment(C);
59
60 // Round up the field offset to the alignment of the field type.
61 uint64_t AlignedNextFieldOffsetInBytes =
62 llvm::RoundUpToAlignment(NextFieldOffsetInBytes, FieldAlignment);
63
64 if (AlignedNextFieldOffsetInBytes > FieldOffsetInBytes) {
65 assert(!Packed && "Alignment is wrong even with a packed struct!");
66
67 // Convert the struct to a packed struct.
68 ConvertStructToPacked();
69
70 AlignedNextFieldOffsetInBytes = NextFieldOffsetInBytes;
71 }
72
73 if (AlignedNextFieldOffsetInBytes < FieldOffsetInBytes) {
74 // We need to append padding.
75 AppendPadding(FieldOffsetInBytes - NextFieldOffsetInBytes);
76
77 assert(NextFieldOffsetInBytes == FieldOffsetInBytes &&
78 "Did not add enough padding!");
79
80 AlignedNextFieldOffsetInBytes = NextFieldOffsetInBytes;
81 }
82
83 // Add the field.
84 Elements.push_back(C);
85 NextFieldOffsetInBytes = AlignedNextFieldOffsetInBytes + getSizeInBytes(C);
86
87 if (Packed)
88 assert(LLVMStructAlignment == 1 && "Packed struct not byte-aligned!");
89 else
90 LLVMStructAlignment = std::max(LLVMStructAlignment, FieldAlignment);
91
92 return true;
93 }
94
95 bool AppendBitField(const FieldDecl *Field, uint64_t FieldOffset,
96 const Expr *InitExpr) {
97 llvm::ConstantInt *CI =
98 cast_or_null<llvm::ConstantInt>(CGM.EmitConstantExpr(InitExpr,
99 Field->getType(),
100 CGF));
101 // FIXME: Can this ever happen?
102 if (!CI)
103 return false;
104
105 if (FieldOffset > NextFieldOffsetInBytes * 8) {
106 // We need to add padding.
107 uint64_t NumBytes =
108 llvm::RoundUpToAlignment(FieldOffset -
109 NextFieldOffsetInBytes * 8, 8) / 8;
110
111 AppendPadding(NumBytes);
112 }
113
114 uint64_t FieldSize =
115 Field->getBitWidth()->EvaluateAsInt(CGM.getContext()).getZExtValue();
116
117 llvm::APInt FieldValue = CI->getValue();
118
119 // Promote the size of FieldValue if necessary
120 // FIXME: This should never occur, but currently it can because initializer
121 // constants are cast to bool, and because clang is not enforcing bitfield
122 // width limits.
123 if (FieldSize > FieldValue.getBitWidth())
124 FieldValue.zext(FieldSize);
125
126 // Truncate the size of FieldValue to the bit field size.
127 if (FieldSize < FieldValue.getBitWidth())
128 FieldValue.trunc(FieldSize);
129
130 if (FieldOffset < NextFieldOffsetInBytes * 8) {
131 // Either part of the field or the entire field can go into the previous
132 // byte.
133 assert(!Elements.empty() && "Elements can't be empty!");
134
135 unsigned BitsInPreviousByte =
136 NextFieldOffsetInBytes * 8 - FieldOffset;
137
138 bool FitsCompletelyInPreviousByte =
139 BitsInPreviousByte >= FieldValue.getBitWidth();
140
141 llvm::APInt Tmp = FieldValue;
142
143 if (!FitsCompletelyInPreviousByte) {
144 unsigned NewFieldWidth = FieldSize - BitsInPreviousByte;
145
146 if (CGM.getTargetData().isBigEndian()) {
147 Tmp = Tmp.lshr(NewFieldWidth);
148 Tmp.trunc(BitsInPreviousByte);
149
150 // We want the remaining high bits.
151 FieldValue.trunc(NewFieldWidth);
152 } else {
153 Tmp.trunc(BitsInPreviousByte);
154
155 // We want the remaining low bits.
156 FieldValue = FieldValue.lshr(BitsInPreviousByte);
157 FieldValue.trunc(NewFieldWidth);
158 }
159 }
160
161 Tmp.zext(8);
162 if (CGM.getTargetData().isBigEndian()) {
163 if (FitsCompletelyInPreviousByte)
164 Tmp = Tmp.shl(BitsInPreviousByte - FieldValue.getBitWidth());
165 } else {
166 Tmp = Tmp.shl(8 - BitsInPreviousByte);
167 }
168
169 // Or in the bits that go into the previous byte.
170 if (llvm::ConstantInt *Val = dyn_cast<llvm::ConstantInt>(Elements.back()))
171 Tmp |= Val->getValue();
172 else
173 assert(isa<llvm::UndefValue>(Elements.back()));
174
175 Elements.back() = llvm::ConstantInt::get(CGM.getLLVMContext(), Tmp);
176
177 if (FitsCompletelyInPreviousByte)
178 return true;
179 }
180
181 while (FieldValue.getBitWidth() > 8) {
182 llvm::APInt Tmp;
183
184 if (CGM.getTargetData().isBigEndian()) {
185 // We want the high bits.
186 Tmp = FieldValue;
187 Tmp = Tmp.lshr(Tmp.getBitWidth() - 8);
188 Tmp.trunc(8);
189 } else {
190 // We want the low bits.
191 Tmp = FieldValue;
192 Tmp.trunc(8);
193
194 FieldValue = FieldValue.lshr(8);
195 }
196
197 Elements.push_back(llvm::ConstantInt::get(CGM.getLLVMContext(), Tmp));
198 NextFieldOffsetInBytes++;
199
200 FieldValue.trunc(FieldValue.getBitWidth() - 8);
201 }
202
203 assert(FieldValue.getBitWidth() > 0 &&
204 "Should have at least one bit left!");
205 assert(FieldValue.getBitWidth() <= 8 &&
206 "Should not have more than a byte left!");
207
208 if (FieldValue.getBitWidth() < 8) {
209 if (CGM.getTargetData().isBigEndian()) {
210 unsigned BitWidth = FieldValue.getBitWidth();
211
212 FieldValue.zext(8);
213 FieldValue = FieldValue << (8 - BitWidth);
214 } else
215 FieldValue.zext(8);
216 }
217
218 // Append the last element.
219 Elements.push_back(llvm::ConstantInt::get(CGM.getLLVMContext(),
220 FieldValue));
221 NextFieldOffsetInBytes++;
222 return true;
223 }
224
225 void AppendPadding(uint64_t NumBytes) {
226 if (!NumBytes)
227 return;
228
229 const llvm::Type *Ty = llvm::Type::getInt8Ty(CGM.getLLVMContext());
230 if (NumBytes > 1)
231 Ty = llvm::ArrayType::get(Ty, NumBytes);
232
233 llvm::Constant *C = llvm::UndefValue::get(Ty);
234 Elements.push_back(C);
235 assert(getAlignment(C) == 1 && "Padding must have 1 byte alignment!");
236
237 NextFieldOffsetInBytes += getSizeInBytes(C);
238 }
239
240 void AppendTailPadding(uint64_t RecordSize) {
241 assert(RecordSize % 8 == 0 && "Invalid record size!");
242
243 uint64_t RecordSizeInBytes = RecordSize / 8;
244 assert(NextFieldOffsetInBytes <= RecordSizeInBytes && "Size mismatch!");
245
246 unsigned NumPadBytes = RecordSizeInBytes - NextFieldOffsetInBytes;
247 AppendPadding(NumPadBytes);
248 }
249
250 void ConvertStructToPacked() {
251 std::vector<llvm::Constant *> PackedElements;
252 uint64_t ElementOffsetInBytes = 0;
253
254 for (unsigned i = 0, e = Elements.size(); i != e; ++i) {
255 llvm::Constant *C = Elements[i];
256
257 unsigned ElementAlign =
258 CGM.getTargetData().getABITypeAlignment(C->getType());
259 uint64_t AlignedElementOffsetInBytes =
260 llvm::RoundUpToAlignment(ElementOffsetInBytes, ElementAlign);
261
262 if (AlignedElementOffsetInBytes > ElementOffsetInBytes) {
263 // We need some padding.
264 uint64_t NumBytes =
265 AlignedElementOffsetInBytes - ElementOffsetInBytes;
266
267 const llvm::Type *Ty = llvm::Type::getInt8Ty(CGM.getLLVMContext());
268 if (NumBytes > 1)
269 Ty = llvm::ArrayType::get(Ty, NumBytes);
270
271 llvm::Constant *Padding = llvm::UndefValue::get(Ty);
272 PackedElements.push_back(Padding);
273 ElementOffsetInBytes += getSizeInBytes(Padding);
274 }
275
276 PackedElements.push_back(C);
277 ElementOffsetInBytes += getSizeInBytes(C);
278 }
279
280 assert(ElementOffsetInBytes == NextFieldOffsetInBytes &&
281 "Packing the struct changed its size!");
282
283 Elements = PackedElements;
284 LLVMStructAlignment = 1;
285 Packed = true;
286 }
287
288 bool Build(InitListExpr *ILE) {
289 RecordDecl *RD = ILE->getType()->getAs<RecordType>()->getDecl();
290 const ASTRecordLayout &Layout = CGM.getContext().getASTRecordLayout(RD);
291
292 unsigned FieldNo = 0;
293 unsigned ElementNo = 0;
294 for (RecordDecl::field_iterator Field = RD->field_begin(),
295 FieldEnd = RD->field_end();
296 ElementNo < ILE->getNumInits() && Field != FieldEnd;
297 ++Field, ++FieldNo) {
298 if (RD->isUnion() && ILE->getInitializedFieldInUnion() != *Field)
299 continue;
300
301 if (Field->isBitField()) {
302 if (!Field->getIdentifier())
303 continue;
304
305 if (!AppendBitField(*Field, Layout.getFieldOffset(FieldNo),
306 ILE->getInit(ElementNo)))
307 return false;
308 } else {
309 if (!AppendField(*Field, Layout.getFieldOffset(FieldNo),
310 ILE->getInit(ElementNo)))
311 return false;
312 }
313
314 ElementNo++;
315 }
316
317 uint64_t LayoutSizeInBytes = Layout.getSize() / 8;
318
319 if (NextFieldOffsetInBytes > LayoutSizeInBytes) {
320 // If the struct is bigger than the size of the record type,
321 // we must have a flexible array member at the end.
322 assert(RD->hasFlexibleArrayMember() &&
323 "Must have flexible array member if struct is bigger than type!");
324
325 // No tail padding is necessary.
326 return true;
327 }
328
329 uint64_t LLVMSizeInBytes = llvm::RoundUpToAlignment(NextFieldOffsetInBytes,
330 LLVMStructAlignment);
331
332 // Check if we need to convert the struct to a packed struct.
333 if (NextFieldOffsetInBytes <= LayoutSizeInBytes &&
334 LLVMSizeInBytes > LayoutSizeInBytes) {
335 assert(!Packed && "Size mismatch!");
336
337 ConvertStructToPacked();
338 assert(NextFieldOffsetInBytes == LayoutSizeInBytes &&
339 "Converting to packed did not help!");
340 }
341
342 // Append tail padding if necessary.
343 AppendTailPadding(Layout.getSize());
344
345 assert(Layout.getSize() / 8 == NextFieldOffsetInBytes &&
346 "Tail padding mismatch!");
347
348 return true;
349 }
350
351 unsigned getAlignment(const llvm::Constant *C) const {
352 if (Packed)
353 return 1;
354
355 return CGM.getTargetData().getABITypeAlignment(C->getType());
356 }
357
358 uint64_t getSizeInBytes(const llvm::Constant *C) const {
359 return CGM.getTargetData().getTypeAllocSize(C->getType());
360 }
361
362public:
363 static llvm::Constant *BuildStruct(CodeGenModule &CGM, CodeGenFunction *CGF,
364 InitListExpr *ILE) {
365 ConstStructBuilder Builder(CGM, CGF);
366
367 if (!Builder.Build(ILE))
368 return 0;
369
370 llvm::Constant *Result =
371 llvm::ConstantStruct::get(CGM.getLLVMContext(),
372 Builder.Elements, Builder.Packed);
373
374 assert(llvm::RoundUpToAlignment(Builder.NextFieldOffsetInBytes,
375 Builder.getAlignment(Result)) ==
376 Builder.getSizeInBytes(Result) && "Size mismatch!");
377
378 return Result;
379 }
380};
381
382class ConstExprEmitter :
383 public StmtVisitor<ConstExprEmitter, llvm::Constant*> {
384 CodeGenModule &CGM;
385 CodeGenFunction *CGF;
386 llvm::LLVMContext &VMContext;
387public:
388 ConstExprEmitter(CodeGenModule &cgm, CodeGenFunction *cgf)
389 : CGM(cgm), CGF(cgf), VMContext(cgm.getLLVMContext()) {
390 }
391
392 //===--------------------------------------------------------------------===//
393 // Visitor Methods
394 //===--------------------------------------------------------------------===//
395
396 llvm::Constant *VisitStmt(Stmt *S) {
397 return 0;
398 }
399
400 llvm::Constant *VisitParenExpr(ParenExpr *PE) {
401 return Visit(PE->getSubExpr());
402 }
403
404 llvm::Constant *VisitCompoundLiteralExpr(CompoundLiteralExpr *E) {
405 return Visit(E->getInitializer());
406 }
407
408 llvm::Constant *EmitMemberFunctionPointer(CXXMethodDecl *MD) {
409 assert(MD->isInstance() && "Member function must not be static!");
410
411 MD = MD->getCanonicalDecl();
412
413 const llvm::Type *PtrDiffTy =
414 CGM.getTypes().ConvertType(CGM.getContext().getPointerDiffType());
415
416 llvm::Constant *Values[2];
417
418 // Get the function pointer (or index if this is a virtual function).
419 if (MD->isVirtual()) {
420 uint64_t Index = CGM.getVtableInfo().getMethodVtableIndex(MD);
421
422 // Itanium C++ ABI 2.3:
423 // For a non-virtual function, this field is a simple function pointer.
424 // For a virtual function, it is 1 plus the virtual table offset
425 // (in bytes) of the function, represented as a ptrdiff_t.
426 Values[0] = llvm::ConstantInt::get(PtrDiffTy, (Index * 8) + 1);
427 } else {
428 const FunctionProtoType *FPT = MD->getType()->getAs<FunctionProtoType>();
429 const llvm::Type *Ty =
430 CGM.getTypes().GetFunctionType(CGM.getTypes().getFunctionInfo(MD),
431 FPT->isVariadic());
432
433 llvm::Constant *FuncPtr = CGM.GetAddrOfFunction(MD, Ty);
434 Values[0] = llvm::ConstantExpr::getPtrToInt(FuncPtr, PtrDiffTy);
435 }
436
437 // The adjustment will always be 0.
438 Values[1] = llvm::ConstantInt::get(PtrDiffTy, 0);
439
440 return llvm::ConstantStruct::get(CGM.getLLVMContext(),
441 Values, 2, /*Packed=*/false);
442 }
443
444 llvm::Constant *VisitUnaryAddrOf(UnaryOperator *E) {
445 if (const MemberPointerType *MPT =
446 E->getType()->getAs<MemberPointerType>()) {
447 QualType T = MPT->getPointeeType();
448 DeclRefExpr *DRE = cast<DeclRefExpr>(E->getSubExpr());
449
450 NamedDecl *ND = DRE->getDecl();
451 if (T->isFunctionProtoType())
452 return EmitMemberFunctionPointer(cast<CXXMethodDecl>(ND));
453
454 // We have a pointer to data member.
455 return CGM.EmitPointerToDataMember(cast<FieldDecl>(ND));
456 }
457
458 return 0;
459 }
460
461 llvm::Constant *VisitBinSub(BinaryOperator *E) {
462 // This must be a pointer/pointer subtraction. This only happens for
463 // address of label.
464 if (!isa<AddrLabelExpr>(E->getLHS()->IgnoreParenNoopCasts(CGM.getContext())) ||
465 !isa<AddrLabelExpr>(E->getRHS()->IgnoreParenNoopCasts(CGM.getContext())))
466 return 0;
467
468 llvm::Constant *LHS = CGM.EmitConstantExpr(E->getLHS(),
469 E->getLHS()->getType(), CGF);
470 llvm::Constant *RHS = CGM.EmitConstantExpr(E->getRHS(),
471 E->getRHS()->getType(), CGF);
472
473 const llvm::Type *ResultType = ConvertType(E->getType());
474 LHS = llvm::ConstantExpr::getPtrToInt(LHS, ResultType);
475 RHS = llvm::ConstantExpr::getPtrToInt(RHS, ResultType);
476
477 // No need to divide by element size, since addr of label is always void*,
478 // which has size 1 in GNUish.
479 return llvm::ConstantExpr::getSub(LHS, RHS);
480 }
481
482 llvm::Constant *VisitCastExpr(CastExpr* E) {
483 switch (E->getCastKind()) {
484 case CastExpr::CK_ToUnion: {
485 // GCC cast to union extension
486 assert(E->getType()->isUnionType() &&
487 "Destination type is not union type!");
488 const llvm::Type *Ty = ConvertType(E->getType());
489 Expr *SubExpr = E->getSubExpr();
490
491 llvm::Constant *C =
492 CGM.EmitConstantExpr(SubExpr, SubExpr->getType(), CGF);
493 if (!C)
494 return 0;
495
496 // Build a struct with the union sub-element as the first member,
497 // and padded to the appropriate size
498 std::vector<llvm::Constant*> Elts;
499 std::vector<const llvm::Type*> Types;
500 Elts.push_back(C);
501 Types.push_back(C->getType());
502 unsigned CurSize = CGM.getTargetData().getTypeAllocSize(C->getType());
503 unsigned TotalSize = CGM.getTargetData().getTypeAllocSize(Ty);
504
505 assert(CurSize <= TotalSize && "Union size mismatch!");
506 if (unsigned NumPadBytes = TotalSize - CurSize) {
507 const llvm::Type *Ty = llvm::Type::getInt8Ty(VMContext);
508 if (NumPadBytes > 1)
509 Ty = llvm::ArrayType::get(Ty, NumPadBytes);
510
511 Elts.push_back(llvm::UndefValue::get(Ty));
512 Types.push_back(Ty);
513 }
514
515 llvm::StructType* STy =
516 llvm::StructType::get(C->getType()->getContext(), Types, false);
517 return llvm::ConstantStruct::get(STy, Elts);
518 }
519 case CastExpr::CK_NullToMemberPointer:
520 return CGM.EmitNullConstant(E->getType());
521
522 case CastExpr::CK_BaseToDerivedMemberPointer: {
523 Expr *SubExpr = E->getSubExpr();
524
525 const MemberPointerType *SrcTy =
526 SubExpr->getType()->getAs<MemberPointerType>();
527 const MemberPointerType *DestTy =
528 E->getType()->getAs<MemberPointerType>();
529
530 const CXXRecordDecl *BaseClass =
531 cast<CXXRecordDecl>(cast<RecordType>(SrcTy->getClass())->getDecl());
532 const CXXRecordDecl *DerivedClass =
533 cast<CXXRecordDecl>(cast<RecordType>(DestTy->getClass())->getDecl());
534
535 if (SrcTy->getPointeeType()->isFunctionProtoType()) {
536 llvm::Constant *C =
537 CGM.EmitConstantExpr(SubExpr, SubExpr->getType(), CGF);
538 if (!C)
539 return 0;
540
541 llvm::ConstantStruct *CS = cast<llvm::ConstantStruct>(C);
542
543 // Check if we need to update the adjustment.
544 if (llvm::Constant *Offset =
545 CGM.GetNonVirtualBaseClassOffset(DerivedClass, BaseClass)) {
546 llvm::Constant *Values[2];
547
548 Values[0] = CS->getOperand(0);
549 Values[1] = llvm::ConstantExpr::getAdd(CS->getOperand(1), Offset);
550 return llvm::ConstantStruct::get(CGM.getLLVMContext(), Values, 2,
551 /*Packed=*/false);
552 }
553
554 return CS;
555 }
556 }
557
558 case CastExpr::CK_BitCast:
559 // This must be a member function pointer cast.
560 return Visit(E->getSubExpr());
561
562 default: {
563 // FIXME: This should be handled by the CK_NoOp cast kind.
564 // Explicit and implicit no-op casts
565 QualType Ty = E->getType(), SubTy = E->getSubExpr()->getType();
566 if (CGM.getContext().hasSameUnqualifiedType(Ty, SubTy))
567 return Visit(E->getSubExpr());
568
569 // Handle integer->integer casts for address-of-label differences.
570 if (Ty->isIntegerType() && SubTy->isIntegerType() &&
571 CGF) {
572 llvm::Value *Src = Visit(E->getSubExpr());
573 if (Src == 0) return 0;
574
575 // Use EmitScalarConversion to perform the conversion.
576 return cast<llvm::Constant>(CGF->EmitScalarConversion(Src, SubTy, Ty));
577 }
578
579 return 0;
580 }
581 }
582 }
583
584 llvm::Constant *VisitCXXDefaultArgExpr(CXXDefaultArgExpr *DAE) {
585 return Visit(DAE->getExpr());
586 }
587
588 llvm::Constant *EmitArrayInitialization(InitListExpr *ILE) {
589 std::vector<llvm::Constant*> Elts;
590 const llvm::ArrayType *AType =
591 cast<llvm::ArrayType>(ConvertType(ILE->getType()));
592 unsigned NumInitElements = ILE->getNumInits();
593 // FIXME: Check for wide strings
594 // FIXME: Check for NumInitElements exactly equal to 1??
595 if (NumInitElements > 0 &&
596 (isa<StringLiteral>(ILE->getInit(0)) ||
597 isa<ObjCEncodeExpr>(ILE->getInit(0))) &&
598 ILE->getType()->getArrayElementTypeNoTypeQual()->isCharType())
599 return Visit(ILE->getInit(0));
600 const llvm::Type *ElemTy = AType->getElementType();
601 unsigned NumElements = AType->getNumElements();
602
603 // Initialising an array requires us to automatically
604 // initialise any elements that have not been initialised explicitly
605 unsigned NumInitableElts = std::min(NumInitElements, NumElements);
606
607 // Copy initializer elements.
608 unsigned i = 0;
609 bool RewriteType = false;
610 for (; i < NumInitableElts; ++i) {
611 Expr *Init = ILE->getInit(i);
612 llvm::Constant *C = CGM.EmitConstantExpr(Init, Init->getType(), CGF);
613 if (!C)
614 return 0;
615 RewriteType |= (C->getType() != ElemTy);
616 Elts.push_back(C);
617 }
618
619 // Initialize remaining array elements.
620 // FIXME: This doesn't handle member pointers correctly!
621 for (; i < NumElements; ++i)
622 Elts.push_back(llvm::Constant::getNullValue(ElemTy));
623
624 if (RewriteType) {
625 // FIXME: Try to avoid packing the array
626 std::vector<const llvm::Type*> Types;
627 for (unsigned i = 0; i < Elts.size(); ++i)
628 Types.push_back(Elts[i]->getType());
629 const llvm::StructType *SType = llvm::StructType::get(AType->getContext(),
630 Types, true);
631 return llvm::ConstantStruct::get(SType, Elts);
632 }
633
634 return llvm::ConstantArray::get(AType, Elts);
635 }
636
637 llvm::Constant *EmitStructInitialization(InitListExpr *ILE) {
638 return ConstStructBuilder::BuildStruct(CGM, CGF, ILE);
639 }
640
641 llvm::Constant *EmitUnionInitialization(InitListExpr *ILE) {
642 return ConstStructBuilder::BuildStruct(CGM, CGF, ILE);
643 }
644
645 llvm::Constant *VisitImplicitValueInitExpr(ImplicitValueInitExpr* E) {
646 return CGM.EmitNullConstant(E->getType());
647 }
648
649 llvm::Constant *VisitInitListExpr(InitListExpr *ILE) {
650 if (ILE->getType()->isScalarType()) {
651 // We have a scalar in braces. Just use the first element.
652 if (ILE->getNumInits() > 0) {
653 Expr *Init = ILE->getInit(0);
654 return CGM.EmitConstantExpr(Init, Init->getType(), CGF);
655 }
656 return CGM.EmitNullConstant(ILE->getType());
657 }
658
659 if (ILE->getType()->isArrayType())
660 return EmitArrayInitialization(ILE);
661
662 if (ILE->getType()->isRecordType())
663 return EmitStructInitialization(ILE);
664
665 if (ILE->getType()->isUnionType())
666 return EmitUnionInitialization(ILE);
667
668 // If ILE was a constant vector, we would have handled it already.
669 if (ILE->getType()->isVectorType())
670 return 0;
671
672 assert(0 && "Unable to handle InitListExpr");
673 // Get rid of control reaches end of void function warning.
674 // Not reached.
675 return 0;
676 }
677
678 llvm::Constant *VisitCXXConstructExpr(CXXConstructExpr *E) {
679 if (!E->getConstructor()->isTrivial())
680 return 0;
681
682 QualType Ty = E->getType();
683
684 // FIXME: We should not have to call getBaseElementType here.
685 const RecordType *RT =
686 CGM.getContext().getBaseElementType(Ty)->getAs<RecordType>();
687 const CXXRecordDecl *RD = cast<CXXRecordDecl>(RT->getDecl());
688
689 // If the class doesn't have a trivial destructor, we can't emit it as a
690 // constant expr.
691 if (!RD->hasTrivialDestructor())
692 return 0;
693
694 // Only copy and default constructors can be trivial.
695
696
697 if (E->getNumArgs()) {
698 assert(E->getNumArgs() == 1 && "trivial ctor with > 1 argument");
699 assert(E->getConstructor()->isCopyConstructor() &&
700 "trivial ctor has argument but isn't a copy ctor");
701
702 Expr *Arg = E->getArg(0);
703 assert(CGM.getContext().hasSameUnqualifiedType(Ty, Arg->getType()) &&
704 "argument to copy ctor is of wrong type");
705
706 return Visit(Arg);
707 }
708
709 return CGM.EmitNullConstant(Ty);
710 }
711
712 llvm::Constant *VisitStringLiteral(StringLiteral *E) {
713 assert(!E->getType()->isPointerType() && "Strings are always arrays");
714
715 // This must be a string initializing an array in a static initializer.
716 // Don't emit it as the address of the string, emit the string data itself
717 // as an inline array.
718 return llvm::ConstantArray::get(VMContext,
719 CGM.GetStringForStringLiteral(E), false);
720 }
721
722 llvm::Constant *VisitObjCEncodeExpr(ObjCEncodeExpr *E) {
723 // This must be an @encode initializing an array in a static initializer.
724 // Don't emit it as the address of the string, emit the string data itself
725 // as an inline array.
726 std::string Str;
727 CGM.getContext().getObjCEncodingForType(E->getEncodedType(), Str);
728 const ConstantArrayType *CAT = cast<ConstantArrayType>(E->getType());
729
730 // Resize the string to the right size, adding zeros at the end, or
731 // truncating as needed.
732 Str.resize(CAT->getSize().getZExtValue(), '\0');
733 return llvm::ConstantArray::get(VMContext, Str, false);
734 }
735
736 llvm::Constant *VisitUnaryExtension(const UnaryOperator *E) {
737 return Visit(E->getSubExpr());
738 }
739
740 // Utility methods
741 const llvm::Type *ConvertType(QualType T) {
742 return CGM.getTypes().ConvertType(T);
743 }
744
745public:
746 llvm::Constant *EmitLValue(Expr *E) {
747 switch (E->getStmtClass()) {
748 default: break;
749 case Expr::CompoundLiteralExprClass: {
750 // Note that due to the nature of compound literals, this is guaranteed
751 // to be the only use of the variable, so we just generate it here.
752 CompoundLiteralExpr *CLE = cast<CompoundLiteralExpr>(E);
753 llvm::Constant* C = Visit(CLE->getInitializer());
754 // FIXME: "Leaked" on failure.
755 if (C)
756 C = new llvm::GlobalVariable(CGM.getModule(), C->getType(),
757 E->getType().isConstant(CGM.getContext()),
758 llvm::GlobalValue::InternalLinkage,
759 C, ".compoundliteral", 0, false,
760 E->getType().getAddressSpace());
761 return C;
762 }
763 case Expr::DeclRefExprClass: {
764 NamedDecl *Decl = cast<DeclRefExpr>(E)->getDecl();
765 if (const FunctionDecl *FD = dyn_cast<FunctionDecl>(Decl))
766 return CGM.GetAddrOfFunction(FD);
767 if (const VarDecl* VD = dyn_cast<VarDecl>(Decl)) {
768 // We can never refer to a variable with local storage.
769 if (!VD->hasLocalStorage()) {
770 if (VD->isFileVarDecl() || VD->hasExternalStorage())
771 return CGM.GetAddrOfGlobalVar(VD);
772 else if (VD->isBlockVarDecl()) {
773 assert(CGF && "Can't access static local vars without CGF");
774 return CGF->GetAddrOfStaticLocalVar(VD);
775 }
776 }
777 }
778 break;
779 }
780 case Expr::StringLiteralClass:
781 return CGM.GetAddrOfConstantStringFromLiteral(cast<StringLiteral>(E));
782 case Expr::ObjCEncodeExprClass:
783 return CGM.GetAddrOfConstantStringFromObjCEncode(cast<ObjCEncodeExpr>(E));
784 case Expr::ObjCStringLiteralClass: {
785 ObjCStringLiteral* SL = cast<ObjCStringLiteral>(E);
786 llvm::Constant *C =
787 CGM.getObjCRuntime().GenerateConstantString(SL->getString());
788 return llvm::ConstantExpr::getBitCast(C, ConvertType(E->getType()));
789 }
790 case Expr::PredefinedExprClass: {
791 unsigned Type = cast<PredefinedExpr>(E)->getIdentType();
792 if (CGF) {
793 LValue Res = CGF->EmitPredefinedFunctionName(Type);
794 return cast<llvm::Constant>(Res.getAddress());
795 } else if (Type == PredefinedExpr::PrettyFunction) {
796 return CGM.GetAddrOfConstantCString("top level", ".tmp");
797 }
798
799 return CGM.GetAddrOfConstantCString("", ".tmp");
800 }
801 case Expr::AddrLabelExprClass: {
802 assert(CGF && "Invalid address of label expression outside function.");
803 llvm::Constant *Ptr =
804 CGF->GetAddrOfLabel(cast<AddrLabelExpr>(E)->getLabel());
805 return llvm::ConstantExpr::getBitCast(Ptr, ConvertType(E->getType()));
806 }
807 case Expr::CallExprClass: {
808 CallExpr* CE = cast<CallExpr>(E);
809 unsigned builtin = CE->isBuiltinCall(CGM.getContext());
810 if (builtin !=
811 Builtin::BI__builtin___CFStringMakeConstantString &&
812 builtin !=
813 Builtin::BI__builtin___NSStringMakeConstantString)
814 break;
815 const Expr *Arg = CE->getArg(0)->IgnoreParenCasts();
816 const StringLiteral *Literal = cast<StringLiteral>(Arg);
817 if (builtin ==
818 Builtin::BI__builtin___NSStringMakeConstantString) {
819 return CGM.getObjCRuntime().GenerateConstantString(Literal);
820 }
821 // FIXME: need to deal with UCN conversion issues.
822 return CGM.GetAddrOfConstantCFString(Literal);
823 }
824 case Expr::BlockExprClass: {
825 std::string FunctionName;
826 if (CGF)
827 FunctionName = CGF->CurFn->getName();
828 else
829 FunctionName = "global";
830
831 return CGM.GetAddrOfGlobalBlock(cast<BlockExpr>(E), FunctionName.c_str());
832 }
833 }
834
835 return 0;
836 }
837};
838
839} // end anonymous namespace.
840
841llvm::Constant *CodeGenModule::EmitConstantExpr(const Expr *E,
842 QualType DestType,
843 CodeGenFunction *CGF) {
844 Expr::EvalResult Result;
845
846 bool Success = false;
847
848 if (DestType->isReferenceType())
849 Success = E->EvaluateAsLValue(Result, Context);
850 else
851 Success = E->Evaluate(Result, Context);
852
853 if (Success && !Result.HasSideEffects) {
854 switch (Result.Val.getKind()) {
855 case APValue::Uninitialized:
856 assert(0 && "Constant expressions should be initialized.");
857 return 0;
858 case APValue::LValue: {
859 const llvm::Type *DestTy = getTypes().ConvertTypeForMem(DestType);
860 llvm::Constant *Offset =
861 llvm::ConstantInt::get(llvm::Type::getInt64Ty(VMContext),
862 Result.Val.getLValueOffset().getQuantity());
863
864 llvm::Constant *C;
865 if (const Expr *LVBase = Result.Val.getLValueBase()) {
866 C = ConstExprEmitter(*this, CGF).EmitLValue(const_cast<Expr*>(LVBase));
867
868 // Apply offset if necessary.
869 if (!Offset->isNullValue()) {
870 const llvm::Type *Type = llvm::Type::getInt8PtrTy(VMContext);
871 llvm::Constant *Casted = llvm::ConstantExpr::getBitCast(C, Type);
872 Casted = llvm::ConstantExpr::getGetElementPtr(Casted, &Offset, 1);
873 C = llvm::ConstantExpr::getBitCast(Casted, C->getType());
874 }
875
876 // Convert to the appropriate type; this could be an lvalue for
877 // an integer.
878 if (isa<llvm::PointerType>(DestTy))
879 return llvm::ConstantExpr::getBitCast(C, DestTy);
880
881 return llvm::ConstantExpr::getPtrToInt(C, DestTy);
882 } else {
883 C = Offset;
884
885 // Convert to the appropriate type; this could be an lvalue for
886 // an integer.
887 if (isa<llvm::PointerType>(DestTy))
888 return llvm::ConstantExpr::getIntToPtr(C, DestTy);
889
890 // If the types don't match this should only be a truncate.
891 if (C->getType() != DestTy)
892 return llvm::ConstantExpr::getTrunc(C, DestTy);
893
894 return C;
895 }
896 }
897 case APValue::Int: {
898 llvm::Constant *C = llvm::ConstantInt::get(VMContext,
899 Result.Val.getInt());
900
901 if (C->getType() == llvm::Type::getInt1Ty(VMContext)) {
902 const llvm::Type *BoolTy = getTypes().ConvertTypeForMem(E->getType());
903 C = llvm::ConstantExpr::getZExt(C, BoolTy);
904 }
905 return C;
906 }
907 case APValue::ComplexInt: {
908 llvm::Constant *Complex[2];
909
910 Complex[0] = llvm::ConstantInt::get(VMContext,
911 Result.Val.getComplexIntReal());
912 Complex[1] = llvm::ConstantInt::get(VMContext,
913 Result.Val.getComplexIntImag());
914
915 // FIXME: the target may want to specify that this is packed.
916 return llvm::ConstantStruct::get(VMContext, Complex, 2, false);
917 }
918 case APValue::Float:
919 return llvm::ConstantFP::get(VMContext, Result.Val.getFloat());
920 case APValue::ComplexFloat: {
921 llvm::Constant *Complex[2];
922
923 Complex[0] = llvm::ConstantFP::get(VMContext,
924 Result.Val.getComplexFloatReal());
925 Complex[1] = llvm::ConstantFP::get(VMContext,
926 Result.Val.getComplexFloatImag());
927
928 // FIXME: the target may want to specify that this is packed.
929 return llvm::ConstantStruct::get(VMContext, Complex, 2, false);
930 }
931 case APValue::Vector: {
932 llvm::SmallVector<llvm::Constant *, 4> Inits;
933 unsigned NumElts = Result.Val.getVectorLength();
934
935 for (unsigned i = 0; i != NumElts; ++i) {
936 APValue &Elt = Result.Val.getVectorElt(i);
937 if (Elt.isInt())
938 Inits.push_back(llvm::ConstantInt::get(VMContext, Elt.getInt()));
939 else
940 Inits.push_back(llvm::ConstantFP::get(VMContext, Elt.getFloat()));
941 }
942 return llvm::ConstantVector::get(&Inits[0], Inits.size());
943 }
944 }
945 }
946
947 llvm::Constant* C = ConstExprEmitter(*this, CGF).Visit(const_cast<Expr*>(E));
948 if (C && C->getType() == llvm::Type::getInt1Ty(VMContext)) {
949 const llvm::Type *BoolTy = getTypes().ConvertTypeForMem(E->getType());
950 C = llvm::ConstantExpr::getZExt(C, BoolTy);
951 }
952 return C;
953}
954
955static bool containsPointerToDataMember(CodeGenTypes &Types, QualType T) {
956 // No need to check for member pointers when not compiling C++.
957 if (!Types.getContext().getLangOptions().CPlusPlus)
958 return false;
959
960 T = Types.getContext().getBaseElementType(T);
961
962 if (const RecordType *RT = T->getAs<RecordType>()) {
963 const CXXRecordDecl *RD = cast<CXXRecordDecl>(RT->getDecl());
964
965 // FIXME: It would be better if there was a way to explicitly compute the
966 // record layout instead of converting to a type.
967 Types.ConvertTagDeclType(RD);
968
969 const CGRecordLayout &Layout = Types.getCGRecordLayout(RD);
970 return Layout.containsPointerToDataMember();
971 }
972
973 if (const MemberPointerType *MPT = T->getAs<MemberPointerType>())
974 return !MPT->getPointeeType()->isFunctionType();
975
976 return false;
977}
978
979llvm::Constant *CodeGenModule::EmitNullConstant(QualType T) {
980 if (!containsPointerToDataMember(getTypes(), T))
981 return llvm::Constant::getNullValue(getTypes().ConvertTypeForMem(T));
982
983 if (const ConstantArrayType *CAT = Context.getAsConstantArrayType(T)) {
984
985 QualType ElementTy = CAT->getElementType();
986
987 llvm::Constant *Element = EmitNullConstant(ElementTy);
988 unsigned NumElements = CAT->getSize().getZExtValue();
989 std::vector<llvm::Constant *> Array(NumElements);
990 for (unsigned i = 0; i != NumElements; ++i)
991 Array[i] = Element;
992
993 const llvm::ArrayType *ATy =
994 cast<llvm::ArrayType>(getTypes().ConvertTypeForMem(T));
995 return llvm::ConstantArray::get(ATy, Array);
996 }
997
998 if (const RecordType *RT = T->getAs<RecordType>()) {
999 const CXXRecordDecl *RD = cast<CXXRecordDecl>(RT->getDecl());
1000 assert(!RD->getNumBases() &&
1001 "FIXME: Handle zero-initializing structs with bases and "
1002 "pointers to data members.");
1003 const llvm::StructType *STy =
1004 cast<llvm::StructType>(getTypes().ConvertTypeForMem(T));
1005 unsigned NumElements = STy->getNumElements();
1006 std::vector<llvm::Constant *> Elements(NumElements);
1007
1008 for (RecordDecl::field_iterator I = RD->field_begin(),
1009 E = RD->field_end(); I != E; ++I) {
1010 const FieldDecl *FD = *I;
1011
1012 unsigned FieldNo = getTypes().getLLVMFieldNo(FD);
1013 Elements[FieldNo] = EmitNullConstant(FD->getType());
1014 }
1015
1016 // Now go through all other fields and zero them out.
1017 for (unsigned i = 0; i != NumElements; ++i) {
1018 if (!Elements[i])
1019 Elements[i] = llvm::Constant::getNullValue(STy->getElementType(i));
1020 }
1021
1022 return llvm::ConstantStruct::get(STy, Elements);
1023 }
1024
1025 assert(!T->getAs<MemberPointerType>()->getPointeeType()->isFunctionType() &&
1026 "Should only see pointers to data members here!");
1027
1028 // Itanium C++ ABI 2.3:
1029 // A NULL pointer is represented as -1.
1030 return llvm::ConstantInt::get(getTypes().ConvertTypeForMem(T), -1,
1031 /*isSigned=*/true);
1032}
1033
1034llvm::Constant *
1035CodeGenModule::EmitPointerToDataMember(const FieldDecl *FD) {
1036
1037 // Itanium C++ ABI 2.3:
1038 // A pointer to data member is an offset from the base address of the class
1039 // object containing it, represented as a ptrdiff_t
1040
1041 const CXXRecordDecl *ClassDecl = cast<CXXRecordDecl>(FD->getParent());
1042 QualType ClassType =
1043 getContext().getTypeDeclType(const_cast<CXXRecordDecl *>(ClassDecl));
1044
1045 const llvm::StructType *ClassLTy =
1046 cast<llvm::StructType>(getTypes().ConvertType(ClassType));
1047
1048 unsigned FieldNo = getTypes().getLLVMFieldNo(FD);
1049 uint64_t Offset =
1050 getTargetData().getStructLayout(ClassLTy)->getElementOffset(FieldNo);
1051
1052 const llvm::Type *PtrDiffTy =
1053 getTypes().ConvertType(getContext().getPointerDiffType());
1054
1055 return llvm::ConstantInt::get(PtrDiffTy, Offset);
1056}