Matt Arsenault | 52ef401 | 2016-07-26 16:45:58 +0000 | [diff] [blame] | 1 | //===-- AMDGPUMachineFunctionInfo.cpp ---------------------------------------=// |
| 2 | // |
| 3 | // The LLVM Compiler Infrastructure |
| 4 | // |
| 5 | // This file is distributed under the University of Illinois Open Source |
| 6 | // License. See LICENSE.TXT for details. |
| 7 | // |
| 8 | //===----------------------------------------------------------------------===// |
| 9 | |
Vincent Lejeune | ace6f73 | 2013-04-01 21:47:53 +0000 | [diff] [blame] | 10 | #include "AMDGPUMachineFunction.h" |
Matt Arsenault | 52ef401 | 2016-07-26 16:45:58 +0000 | [diff] [blame] | 11 | #include "AMDGPUSubtarget.h" |
Stanislav Mekhanoshin | 1c53842 | 2018-05-25 17:25:12 +0000 | [diff] [blame] | 12 | #include "AMDGPUPerfHintAnalysis.h" |
| 13 | #include "llvm/CodeGen/MachineModuleInfo.h" |
Matt Arsenault | e935f05 | 2016-06-18 05:15:53 +0000 | [diff] [blame] | 14 | |
Craig Topper | 8fc4096 | 2013-07-17 00:31:35 +0000 | [diff] [blame] | 15 | using namespace llvm; |
Vincent Lejeune | ace6f73 | 2013-04-01 21:47:53 +0000 | [diff] [blame] | 16 | |
Vincent Lejeune | ace6f73 | 2013-04-01 21:47:53 +0000 | [diff] [blame] | 17 | AMDGPUMachineFunction::AMDGPUMachineFunction(const MachineFunction &MF) : |
Matt Arsenault | 762af96 | 2014-07-13 03:06:39 +0000 | [diff] [blame] | 18 | MachineFunctionInfo(), |
Matt Arsenault | 52ef401 | 2016-07-26 16:45:58 +0000 | [diff] [blame] | 19 | LocalMemoryObjects(), |
Matt Arsenault | 75e7192 | 2018-06-28 10:18:55 +0000 | [diff] [blame^] | 20 | ExplicitKernArgSize(0), |
Matt Arsenault | e935f05 | 2016-06-18 05:15:53 +0000 | [diff] [blame] | 21 | MaxKernArgAlign(0), |
Matt Arsenault | 3f98140 | 2014-09-15 15:41:53 +0000 | [diff] [blame] | 22 | LDSSize(0), |
Matthias Braun | f1caa28 | 2017-12-15 22:22:58 +0000 | [diff] [blame] | 23 | IsEntryFunction(AMDGPU::isEntryFunctionCC(MF.getFunction().getCallingConv())), |
Stanislav Mekhanoshin | 1c53842 | 2018-05-25 17:25:12 +0000 | [diff] [blame] | 24 | NoSignedZerosFPMath(MF.getTarget().Options.NoSignedZerosFPMath), |
| 25 | MemoryBound(false), |
| 26 | WaveLimiter(false) { |
Matt Arsenault | 52ef401 | 2016-07-26 16:45:58 +0000 | [diff] [blame] | 27 | // FIXME: Should initialize KernArgSize based on ExplicitKernelArgOffset, |
| 28 | // except reserved size is not correctly aligned. |
Stanislav Mekhanoshin | 1c53842 | 2018-05-25 17:25:12 +0000 | [diff] [blame] | 29 | |
| 30 | if (auto *Resolver = MF.getMMI().getResolver()) { |
| 31 | if (AMDGPUPerfHintAnalysis *PHA = static_cast<AMDGPUPerfHintAnalysis*>( |
| 32 | Resolver->getAnalysisIfAvailable(&AMDGPUPerfHintAnalysisID, true))) { |
| 33 | MemoryBound = PHA->isMemoryBound(&MF.getFunction()); |
| 34 | WaveLimiter = PHA->needsWaveLimiter(&MF.getFunction()); |
| 35 | } |
| 36 | } |
Nikolay Haustov | beb24f5 | 2016-07-01 10:00:58 +0000 | [diff] [blame] | 37 | } |
| 38 | |
Matt Arsenault | 52ef401 | 2016-07-26 16:45:58 +0000 | [diff] [blame] | 39 | unsigned AMDGPUMachineFunction::allocateLDSGlobal(const DataLayout &DL, |
| 40 | const GlobalValue &GV) { |
| 41 | auto Entry = LocalMemoryObjects.insert(std::make_pair(&GV, 0)); |
| 42 | if (!Entry.second) |
| 43 | return Entry.first->second; |
| 44 | |
| 45 | unsigned Align = GV.getAlignment(); |
| 46 | if (Align == 0) |
| 47 | Align = DL.getABITypeAlignment(GV.getValueType()); |
| 48 | |
| 49 | /// TODO: We should sort these to minimize wasted space due to alignment |
| 50 | /// padding. Currently the padding is decided by the first encountered use |
| 51 | /// during lowering. |
| 52 | unsigned Offset = LDSSize = alignTo(LDSSize, Align); |
| 53 | |
| 54 | Entry.first->second = Offset; |
| 55 | LDSSize += DL.getTypeAllocSize(GV.getValueType()); |
| 56 | |
| 57 | return Offset; |
Nikolay Haustov | beb24f5 | 2016-07-01 10:00:58 +0000 | [diff] [blame] | 58 | } |