| //===-- AMDGPULowerKernelAttributes.cpp ------------------------------------------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| /// \file This pass does attempts to make use of reqd_work_group_size metadata |
| /// to eliminate loads from the dispatch packet and to constant fold OpenCL |
| /// get_local_size-like functions. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "AMDGPU.h" |
| #include "AMDGPUTargetMachine.h" |
| #include "llvm/Analysis/ValueTracking.h" |
| #include "llvm/CodeGen/Passes.h" |
| #include "llvm/CodeGen/TargetPassConfig.h" |
| #include "llvm/IR/Constants.h" |
| #include "llvm/IR/Function.h" |
| #include "llvm/IR/Instructions.h" |
| #include "llvm/IR/PatternMatch.h" |
| #include "llvm/Pass.h" |
| |
| #define DEBUG_TYPE "amdgpu-lower-kernel-attributes" |
| |
| using namespace llvm; |
| |
| namespace { |
| |
| // Field offsets in hsa_kernel_dispatch_packet_t. |
| enum DispatchPackedOffsets { |
| WORKGROUP_SIZE_X = 4, |
| WORKGROUP_SIZE_Y = 6, |
| WORKGROUP_SIZE_Z = 8, |
| |
| GRID_SIZE_X = 12, |
| GRID_SIZE_Y = 16, |
| GRID_SIZE_Z = 20 |
| }; |
| |
| class AMDGPULowerKernelAttributes : public ModulePass { |
| Module *Mod = nullptr; |
| |
| public: |
| static char ID; |
| |
| AMDGPULowerKernelAttributes() : ModulePass(ID) {} |
| |
| bool processUse(CallInst *CI); |
| |
| bool doInitialization(Module &M) override; |
| bool runOnModule(Module &M) override; |
| |
| StringRef getPassName() const override { |
| return "AMDGPU Kernel Attributes"; |
| } |
| |
| void getAnalysisUsage(AnalysisUsage &AU) const override { |
| AU.setPreservesAll(); |
| } |
| }; |
| |
| } // end anonymous namespace |
| |
| bool AMDGPULowerKernelAttributes::doInitialization(Module &M) { |
| Mod = &M; |
| return false; |
| } |
| |
| bool AMDGPULowerKernelAttributes::processUse(CallInst *CI) { |
| Function *F = CI->getParent()->getParent(); |
| |
| auto MD = F->getMetadata("reqd_work_group_size"); |
| const bool HasReqdWorkGroupSize = MD && MD->getNumOperands() == 3; |
| |
| const bool HasUniformWorkGroupSize = |
| F->getFnAttribute("uniform-work-group-size").getValueAsString() == "true"; |
| |
| if (!HasReqdWorkGroupSize && !HasUniformWorkGroupSize) |
| return false; |
| |
| Value *WorkGroupSizeX = nullptr; |
| Value *WorkGroupSizeY = nullptr; |
| Value *WorkGroupSizeZ = nullptr; |
| |
| Value *GridSizeX = nullptr; |
| Value *GridSizeY = nullptr; |
| Value *GridSizeZ = nullptr; |
| |
| const DataLayout &DL = Mod->getDataLayout(); |
| |
| // We expect to see several GEP users, casted to the appropriate type and |
| // loaded. |
| for (User *U : CI->users()) { |
| if (!U->hasOneUse()) |
| continue; |
| |
| int64_t Offset = 0; |
| if (GetPointerBaseWithConstantOffset(U, Offset, DL) != CI) |
| continue; |
| |
| auto *BCI = dyn_cast<BitCastInst>(*U->user_begin()); |
| if (!BCI || !BCI->hasOneUse()) |
| continue; |
| |
| auto *Load = dyn_cast<LoadInst>(*BCI->user_begin()); |
| if (!Load || !Load->isSimple()) |
| continue; |
| |
| unsigned LoadSize = DL.getTypeStoreSize(Load->getType()); |
| |
| // TODO: Handle merged loads. |
| switch (Offset) { |
| case WORKGROUP_SIZE_X: |
| if (LoadSize == 2) |
| WorkGroupSizeX = Load; |
| break; |
| case WORKGROUP_SIZE_Y: |
| if (LoadSize == 2) |
| WorkGroupSizeY = Load; |
| break; |
| case WORKGROUP_SIZE_Z: |
| if (LoadSize == 2) |
| WorkGroupSizeZ = Load; |
| break; |
| case GRID_SIZE_X: |
| if (LoadSize == 4) |
| GridSizeX = Load; |
| break; |
| case GRID_SIZE_Y: |
| if (LoadSize == 4) |
| GridSizeY = Load; |
| break; |
| case GRID_SIZE_Z: |
| if (LoadSize == 4) |
| GridSizeZ = Load; |
| break; |
| default: |
| break; |
| } |
| } |
| |
| // Pattern match the code used to handle partial workgroup dispatches in the |
| // library implementation of get_local_size, so the entire function can be |
| // constant folded with a known group size. |
| // |
| // uint r = grid_size - group_id * group_size; |
| // get_local_size = (r < group_size) ? r : group_size; |
| // |
| // If we have uniform-work-group-size (which is the default in OpenCL 1.2), |
| // the grid_size is required to be a multiple of group_size). In this case: |
| // |
| // grid_size - (group_id * group_size) < group_size |
| // -> |
| // grid_size < group_size + (group_id * group_size) |
| // |
| // (grid_size / group_size) < 1 + group_id |
| // |
| // grid_size / group_size is at least 1, so we can conclude the select |
| // condition is false (except for group_id == 0, where the select result is |
| // the same). |
| |
| bool MadeChange = false; |
| Value *WorkGroupSizes[3] = { WorkGroupSizeX, WorkGroupSizeY, WorkGroupSizeZ }; |
| Value *GridSizes[3] = { GridSizeX, GridSizeY, GridSizeZ }; |
| |
| for (int I = 0; HasUniformWorkGroupSize && I < 3; ++I) { |
| Value *GroupSize = WorkGroupSizes[I]; |
| Value *GridSize = GridSizes[I]; |
| if (!GroupSize || !GridSize) |
| continue; |
| |
| for (User *U : GroupSize->users()) { |
| auto *ZextGroupSize = dyn_cast<ZExtInst>(U); |
| if (!ZextGroupSize) |
| continue; |
| |
| for (User *ZextUser : ZextGroupSize->users()) { |
| auto *SI = dyn_cast<SelectInst>(ZextUser); |
| if (!SI) |
| continue; |
| |
| using namespace llvm::PatternMatch; |
| auto GroupIDIntrin = I == 0 ? |
| m_Intrinsic<Intrinsic::amdgcn_workgroup_id_x>() : |
| (I == 1 ? m_Intrinsic<Intrinsic::amdgcn_workgroup_id_y>() : |
| m_Intrinsic<Intrinsic::amdgcn_workgroup_id_z>()); |
| |
| auto SubExpr = m_Sub(m_Specific(GridSize), |
| m_Mul(GroupIDIntrin, m_Specific(ZextGroupSize))); |
| |
| ICmpInst::Predicate Pred; |
| if (match(SI, |
| m_Select(m_ICmp(Pred, SubExpr, m_Specific(ZextGroupSize)), |
| SubExpr, |
| m_Specific(ZextGroupSize))) && |
| Pred == ICmpInst::ICMP_ULT) { |
| if (HasReqdWorkGroupSize) { |
| ConstantInt *KnownSize |
| = mdconst::extract<ConstantInt>(MD->getOperand(I)); |
| SI->replaceAllUsesWith(ConstantExpr::getIntegerCast(KnownSize, |
| SI->getType(), |
| false)); |
| } else { |
| SI->replaceAllUsesWith(ZextGroupSize); |
| } |
| |
| MadeChange = true; |
| } |
| } |
| } |
| } |
| |
| if (!HasReqdWorkGroupSize) |
| return MadeChange; |
| |
| // Eliminate any other loads we can from the dispatch packet. |
| for (int I = 0; I < 3; ++I) { |
| Value *GroupSize = WorkGroupSizes[I]; |
| if (!GroupSize) |
| continue; |
| |
| ConstantInt *KnownSize = mdconst::extract<ConstantInt>(MD->getOperand(I)); |
| GroupSize->replaceAllUsesWith( |
| ConstantExpr::getIntegerCast(KnownSize, |
| GroupSize->getType(), |
| false)); |
| MadeChange = true; |
| } |
| |
| return MadeChange; |
| } |
| |
| // TODO: Move makeLIDRangeMetadata usage into here. Seem to not get |
| // TargetPassConfig for subtarget. |
| bool AMDGPULowerKernelAttributes::runOnModule(Module &M) { |
| StringRef DispatchPtrName |
| = Intrinsic::getName(Intrinsic::amdgcn_dispatch_ptr); |
| |
| Function *DispatchPtr = Mod->getFunction(DispatchPtrName); |
| if (!DispatchPtr) // Dispatch ptr not used. |
| return false; |
| |
| bool MadeChange = false; |
| |
| SmallPtrSet<Instruction *, 4> HandledUses; |
| for (auto *U : DispatchPtr->users()) { |
| CallInst *CI = cast<CallInst>(U); |
| if (HandledUses.insert(CI).second) { |
| if (processUse(CI)) |
| MadeChange = true; |
| } |
| } |
| |
| return MadeChange; |
| } |
| |
| INITIALIZE_PASS_BEGIN(AMDGPULowerKernelAttributes, DEBUG_TYPE, |
| "AMDGPU IR optimizations", false, false) |
| INITIALIZE_PASS_END(AMDGPULowerKernelAttributes, DEBUG_TYPE, "AMDGPU IR optimizations", |
| false, false) |
| |
| char AMDGPULowerKernelAttributes::ID = 0; |
| |
| ModulePass *llvm::createAMDGPULowerKernelAttributesPass() { |
| return new AMDGPULowerKernelAttributes(); |
| } |