| //===- SIMemoryLegalizer.cpp ----------------------------------------------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| /// \file |
| /// Memory legalizer - implements memory model. More information can be |
| /// found here: |
| /// http://llvm.org/docs/AMDGPUUsage.html#memory-model |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "AMDGPU.h" |
| #include "AMDGPUMachineModuleInfo.h" |
| #include "AMDGPUSubtarget.h" |
| #include "SIDefines.h" |
| #include "SIInstrInfo.h" |
| #include "MCTargetDesc/AMDGPUMCTargetDesc.h" |
| #include "Utils/AMDGPUBaseInfo.h" |
| #include "llvm/ADT/BitmaskEnum.h" |
| #include "llvm/ADT/None.h" |
| #include "llvm/ADT/Optional.h" |
| #include "llvm/CodeGen/MachineBasicBlock.h" |
| #include "llvm/CodeGen/MachineFunction.h" |
| #include "llvm/CodeGen/MachineFunctionPass.h" |
| #include "llvm/CodeGen/MachineInstrBuilder.h" |
| #include "llvm/CodeGen/MachineMemOperand.h" |
| #include "llvm/CodeGen/MachineModuleInfo.h" |
| #include "llvm/CodeGen/MachineOperand.h" |
| #include "llvm/IR/DebugLoc.h" |
| #include "llvm/IR/DiagnosticInfo.h" |
| #include "llvm/IR/Function.h" |
| #include "llvm/IR/LLVMContext.h" |
| #include "llvm/MC/MCInstrDesc.h" |
| #include "llvm/Pass.h" |
| #include "llvm/Support/AtomicOrdering.h" |
| #include "llvm/Support/MathExtras.h" |
| #include <cassert> |
| #include <list> |
| |
| using namespace llvm; |
| using namespace llvm::AMDGPU; |
| |
| #define DEBUG_TYPE "si-memory-legalizer" |
| #define PASS_NAME "SI Memory Legalizer" |
| |
| namespace { |
| |
| LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE(); |
| |
| /// Memory operation flags. Can be ORed together. |
| enum class SIMemOp { |
| NONE = 0u, |
| LOAD = 1u << 0, |
| STORE = 1u << 1, |
| LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ STORE) |
| }; |
| |
| /// Position to insert a new instruction relative to an existing |
| /// instruction. |
| enum class Position { |
| BEFORE, |
| AFTER |
| }; |
| |
| /// The atomic synchronization scopes supported by the AMDGPU target. |
| enum class SIAtomicScope { |
| NONE, |
| SINGLETHREAD, |
| WAVEFRONT, |
| WORKGROUP, |
| AGENT, |
| SYSTEM |
| }; |
| |
| /// The distinct address spaces supported by the AMDGPU target for |
| /// atomic memory operation. Can be ORed toether. |
| enum class SIAtomicAddrSpace { |
| NONE = 0u, |
| GLOBAL = 1u << 0, |
| LDS = 1u << 1, |
| SCRATCH = 1u << 2, |
| GDS = 1u << 3, |
| OTHER = 1u << 4, |
| |
| /// The address spaces that can be accessed by a FLAT instruction. |
| FLAT = GLOBAL | LDS | SCRATCH, |
| |
| /// The address spaces that support atomic instructions. |
| ATOMIC = GLOBAL | LDS | SCRATCH | GDS, |
| |
| /// All address spaces. |
| ALL = GLOBAL | LDS | SCRATCH | GDS | OTHER, |
| |
| LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ ALL) |
| }; |
| |
| /// Sets named bit \p BitName to "true" if present in instruction \p MI. |
| /// \returns Returns true if \p MI is modified, false otherwise. |
| template <uint16_t BitName> |
| bool enableNamedBit(const MachineBasicBlock::iterator &MI) { |
| int BitIdx = AMDGPU::getNamedOperandIdx(MI->getOpcode(), BitName); |
| if (BitIdx == -1) |
| return false; |
| |
| MachineOperand &Bit = MI->getOperand(BitIdx); |
| if (Bit.getImm() != 0) |
| return false; |
| |
| Bit.setImm(1); |
| return true; |
| } |
| |
| class SIMemOpInfo final { |
| private: |
| |
| friend class SIMemOpAccess; |
| |
| AtomicOrdering Ordering = AtomicOrdering::NotAtomic; |
| AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; |
| SIAtomicScope Scope = SIAtomicScope::SYSTEM; |
| SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; |
| SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; |
| bool IsCrossAddressSpaceOrdering = false; |
| bool IsNonTemporal = false; |
| |
| SIMemOpInfo(AtomicOrdering Ordering = AtomicOrdering::SequentiallyConsistent, |
| SIAtomicScope Scope = SIAtomicScope::SYSTEM, |
| SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC, |
| SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL, |
| bool IsCrossAddressSpaceOrdering = true, |
| AtomicOrdering FailureOrdering = |
| AtomicOrdering::SequentiallyConsistent, |
| bool IsNonTemporal = false) |
| : Ordering(Ordering), FailureOrdering(FailureOrdering), |
| Scope(Scope), OrderingAddrSpace(OrderingAddrSpace), |
| InstrAddrSpace(InstrAddrSpace), |
| IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering), |
| IsNonTemporal(IsNonTemporal) { |
| // There is also no cross address space ordering if the ordering |
| // address space is the same as the instruction address space and |
| // only contains a single address space. |
| if ((OrderingAddrSpace == InstrAddrSpace) && |
| isPowerOf2_32(uint32_t(InstrAddrSpace))) |
| IsCrossAddressSpaceOrdering = false; |
| } |
| |
| public: |
| /// \returns Atomic synchronization scope of the machine instruction used to |
| /// create this SIMemOpInfo. |
| SIAtomicScope getScope() const { |
| return Scope; |
| } |
| |
| /// \returns Ordering constraint of the machine instruction used to |
| /// create this SIMemOpInfo. |
| AtomicOrdering getOrdering() const { |
| return Ordering; |
| } |
| |
| /// \returns Failure ordering constraint of the machine instruction used to |
| /// create this SIMemOpInfo. |
| AtomicOrdering getFailureOrdering() const { |
| return FailureOrdering; |
| } |
| |
| /// \returns The address spaces be accessed by the machine |
| /// instruction used to create this SiMemOpInfo. |
| SIAtomicAddrSpace getInstrAddrSpace() const { |
| return InstrAddrSpace; |
| } |
| |
| /// \returns The address spaces that must be ordered by the machine |
| /// instruction used to create this SiMemOpInfo. |
| SIAtomicAddrSpace getOrderingAddrSpace() const { |
| return OrderingAddrSpace; |
| } |
| |
| /// \returns Return true iff memory ordering of operations on |
| /// different address spaces is required. |
| bool getIsCrossAddressSpaceOrdering() const { |
| return IsCrossAddressSpaceOrdering; |
| } |
| |
| /// \returns True if memory access of the machine instruction used to |
| /// create this SIMemOpInfo is non-temporal, false otherwise. |
| bool isNonTemporal() const { |
| return IsNonTemporal; |
| } |
| |
| /// \returns True if ordering constraint of the machine instruction used to |
| /// create this SIMemOpInfo is unordered or higher, false otherwise. |
| bool isAtomic() const { |
| return Ordering != AtomicOrdering::NotAtomic; |
| } |
| |
| }; |
| |
| class SIMemOpAccess final { |
| private: |
| |
| AMDGPUAS SIAddrSpaceInfo; |
| AMDGPUMachineModuleInfo *MMI = nullptr; |
| |
| /// Reports unsupported message \p Msg for \p MI to LLVM context. |
| void reportUnsupported(const MachineBasicBlock::iterator &MI, |
| const char *Msg) const; |
| |
| /// Inspects the target synchonization scope \p SSID and determines |
| /// the SI atomic scope it corresponds to, the address spaces it |
| /// covers, and whether the memory ordering applies between address |
| /// spaces. |
| Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> |
| toSIAtomicScope(SyncScope::ID SSID, SIAtomicAddrSpace InstrScope) const; |
| |
| /// \return Return a bit set of the address spaces accessed by \p AS. |
| SIAtomicAddrSpace toSIAtomicAddrSpace(unsigned AS) const; |
| |
| /// \returns Info constructed from \p MI, which has at least machine memory |
| /// operand. |
| Optional<SIMemOpInfo> constructFromMIWithMMO( |
| const MachineBasicBlock::iterator &MI) const; |
| |
| public: |
| /// Construct class to support accessing the machine memory operands |
| /// of instructions in the machine function \p MF. |
| SIMemOpAccess(MachineFunction &MF); |
| |
| /// \returns Load info if \p MI is a load operation, "None" otherwise. |
| Optional<SIMemOpInfo> getLoadInfo( |
| const MachineBasicBlock::iterator &MI) const; |
| |
| /// \returns Store info if \p MI is a store operation, "None" otherwise. |
| Optional<SIMemOpInfo> getStoreInfo( |
| const MachineBasicBlock::iterator &MI) const; |
| |
| /// \returns Atomic fence info if \p MI is an atomic fence operation, |
| /// "None" otherwise. |
| Optional<SIMemOpInfo> getAtomicFenceInfo( |
| const MachineBasicBlock::iterator &MI) const; |
| |
| /// \returns Atomic cmpxchg/rmw info if \p MI is an atomic cmpxchg or |
| /// rmw operation, "None" otherwise. |
| Optional<SIMemOpInfo> getAtomicCmpxchgOrRmwInfo( |
| const MachineBasicBlock::iterator &MI) const; |
| }; |
| |
| class SICacheControl { |
| protected: |
| |
| /// Instruction info. |
| const SIInstrInfo *TII = nullptr; |
| |
| IsaInfo::IsaVersion IV; |
| |
| SICacheControl(const GCNSubtarget &ST); |
| |
| public: |
| |
| /// Create a cache control for the subtarget \p ST. |
| static std::unique_ptr<SICacheControl> create(const GCNSubtarget &ST); |
| |
| /// Update \p MI memory load instruction to bypass any caches up to |
| /// the \p Scope memory scope for address spaces \p |
| /// AddrSpace. Return true iff the instruction was modified. |
| virtual bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace) const = 0; |
| |
| /// Update \p MI memory instruction to indicate it is |
| /// nontemporal. Return true iff the instruction was modified. |
| virtual bool enableNonTemporal(const MachineBasicBlock::iterator &MI) |
| const = 0; |
| |
| /// Inserts any necessary instructions at position \p Pos relative |
| /// to instruction \p MI to ensure any caches associated with |
| /// address spaces \p AddrSpace for memory scopes up to memory scope |
| /// \p Scope are invalidated. Returns true iff any instructions |
| /// inserted. |
| virtual bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| Position Pos) const = 0; |
| |
| /// Inserts any necessary instructions at position \p Pos relative |
| /// to instruction \p MI to ensure memory instructions of kind \p Op |
| /// associated with address spaces \p AddrSpace have completed as |
| /// observed by other memory instructions executing in memory scope |
| /// \p Scope. \p IsCrossAddrSpaceOrdering indicates if the memory |
| /// ordering is between address spaces. Returns true iff any |
| /// instructions inserted. |
| virtual bool insertWait(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| SIMemOp Op, |
| bool IsCrossAddrSpaceOrdering, |
| Position Pos) const = 0; |
| |
| /// Virtual destructor to allow derivations to be deleted. |
| virtual ~SICacheControl() = default; |
| |
| }; |
| |
| class SIGfx6CacheControl : public SICacheControl { |
| protected: |
| |
| /// Sets GLC bit to "true" if present in \p MI. Returns true if \p MI |
| /// is modified, false otherwise. |
| bool enableGLCBit(const MachineBasicBlock::iterator &MI) const { |
| return enableNamedBit<AMDGPU::OpName::glc>(MI); |
| } |
| |
| /// Sets SLC bit to "true" if present in \p MI. Returns true if \p MI |
| /// is modified, false otherwise. |
| bool enableSLCBit(const MachineBasicBlock::iterator &MI) const { |
| return enableNamedBit<AMDGPU::OpName::slc>(MI); |
| } |
| |
| public: |
| |
| SIGfx6CacheControl(const GCNSubtarget &ST) : SICacheControl(ST) {}; |
| |
| bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace) const override; |
| |
| bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override; |
| |
| bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| Position Pos) const override; |
| |
| bool insertWait(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| SIMemOp Op, |
| bool IsCrossAddrSpaceOrdering, |
| Position Pos) const override; |
| }; |
| |
| class SIGfx7CacheControl : public SIGfx6CacheControl { |
| public: |
| |
| SIGfx7CacheControl(const GCNSubtarget &ST) : SIGfx6CacheControl(ST) {}; |
| |
| bool insertCacheInvalidate(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| Position Pos) const override; |
| |
| }; |
| |
| class SIMemoryLegalizer final : public MachineFunctionPass { |
| private: |
| |
| /// Cache Control. |
| std::unique_ptr<SICacheControl> CC = nullptr; |
| |
| /// List of atomic pseudo instructions. |
| std::list<MachineBasicBlock::iterator> AtomicPseudoMIs; |
| |
| /// Return true iff instruction \p MI is a atomic instruction that |
| /// returns a result. |
| bool isAtomicRet(const MachineInstr &MI) const { |
| return AMDGPU::getAtomicNoRetOp(MI.getOpcode()) != -1; |
| } |
| |
| /// Removes all processed atomic pseudo instructions from the current |
| /// function. Returns true if current function is modified, false otherwise. |
| bool removeAtomicPseudoMIs(); |
| |
| /// Expands load operation \p MI. Returns true if instructions are |
| /// added/deleted or \p MI is modified, false otherwise. |
| bool expandLoad(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI); |
| /// Expands store operation \p MI. Returns true if instructions are |
| /// added/deleted or \p MI is modified, false otherwise. |
| bool expandStore(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI); |
| /// Expands atomic fence operation \p MI. Returns true if |
| /// instructions are added/deleted or \p MI is modified, false otherwise. |
| bool expandAtomicFence(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI); |
| /// Expands atomic cmpxchg or rmw operation \p MI. Returns true if |
| /// instructions are added/deleted or \p MI is modified, false otherwise. |
| bool expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI); |
| |
| public: |
| static char ID; |
| |
| SIMemoryLegalizer() : MachineFunctionPass(ID) {} |
| |
| void getAnalysisUsage(AnalysisUsage &AU) const override { |
| AU.setPreservesCFG(); |
| MachineFunctionPass::getAnalysisUsage(AU); |
| } |
| |
| StringRef getPassName() const override { |
| return PASS_NAME; |
| } |
| |
| bool runOnMachineFunction(MachineFunction &MF) override; |
| }; |
| |
| } // end namespace anonymous |
| |
| void SIMemOpAccess::reportUnsupported(const MachineBasicBlock::iterator &MI, |
| const char *Msg) const { |
| const Function &Func = MI->getParent()->getParent()->getFunction(); |
| DiagnosticInfoUnsupported Diag(Func, Msg, MI->getDebugLoc()); |
| Func.getContext().diagnose(Diag); |
| } |
| |
| Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>> |
| SIMemOpAccess::toSIAtomicScope(SyncScope::ID SSID, |
| SIAtomicAddrSpace InstrScope) const { |
| /// TODO: For now assume OpenCL memory model which treats each |
| /// address space as having a separate happens-before relation, and |
| /// so an instruction only has ordering with respect to the address |
| /// space it accesses, and if it accesses multiple address spaces it |
| /// does not require ordering of operations in different address |
| /// spaces. |
| if (SSID == SyncScope::System) |
| return std::make_tuple(SIAtomicScope::SYSTEM, |
| SIAtomicAddrSpace::ATOMIC & InstrScope, |
| false); |
| if (SSID == MMI->getAgentSSID()) |
| return std::make_tuple(SIAtomicScope::AGENT, |
| SIAtomicAddrSpace::ATOMIC & InstrScope, |
| false); |
| if (SSID == MMI->getWorkgroupSSID()) |
| return std::make_tuple(SIAtomicScope::WORKGROUP, |
| SIAtomicAddrSpace::ATOMIC & InstrScope, |
| false); |
| if (SSID == MMI->getWavefrontSSID()) |
| return std::make_tuple(SIAtomicScope::WAVEFRONT, |
| SIAtomicAddrSpace::ATOMIC & InstrScope, |
| false); |
| if (SSID == SyncScope::SingleThread) |
| return std::make_tuple(SIAtomicScope::SINGLETHREAD, |
| SIAtomicAddrSpace::ATOMIC & InstrScope, |
| false); |
| /// TODO: To support HSA Memory Model need to add additional memory |
| /// scopes that specify that do require cross address space |
| /// ordering. |
| return None; |
| } |
| |
| SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(unsigned AS) const { |
| if (AS == SIAddrSpaceInfo.FLAT_ADDRESS) |
| return SIAtomicAddrSpace::FLAT; |
| if (AS == SIAddrSpaceInfo.GLOBAL_ADDRESS) |
| return SIAtomicAddrSpace::GLOBAL; |
| if (AS == SIAddrSpaceInfo.LOCAL_ADDRESS) |
| return SIAtomicAddrSpace::LDS; |
| if (AS == SIAddrSpaceInfo.PRIVATE_ADDRESS) |
| return SIAtomicAddrSpace::SCRATCH; |
| if (AS == SIAddrSpaceInfo.REGION_ADDRESS) |
| return SIAtomicAddrSpace::GDS; |
| |
| return SIAtomicAddrSpace::OTHER; |
| } |
| |
| SIMemOpAccess::SIMemOpAccess(MachineFunction &MF) { |
| SIAddrSpaceInfo = getAMDGPUAS(MF.getTarget()); |
| MMI = &MF.getMMI().getObjFileInfo<AMDGPUMachineModuleInfo>(); |
| } |
| |
| Optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->getNumMemOperands() > 0); |
| |
| SyncScope::ID SSID = SyncScope::SingleThread; |
| AtomicOrdering Ordering = AtomicOrdering::NotAtomic; |
| AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic; |
| SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE; |
| bool IsNonTemporal = true; |
| |
| // Validator should check whether or not MMOs cover the entire set of |
| // locations accessed by the memory instruction. |
| for (const auto &MMO : MI->memoperands()) { |
| IsNonTemporal &= MMO->isNonTemporal(); |
| InstrAddrSpace |= |
| toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace()); |
| AtomicOrdering OpOrdering = MMO->getOrdering(); |
| if (OpOrdering != AtomicOrdering::NotAtomic) { |
| const auto &IsSyncScopeInclusion = |
| MMI->isSyncScopeInclusion(SSID, MMO->getSyncScopeID()); |
| if (!IsSyncScopeInclusion) { |
| reportUnsupported(MI, |
| "Unsupported non-inclusive atomic synchronization scope"); |
| return None; |
| } |
| |
| SSID = IsSyncScopeInclusion.getValue() ? SSID : MMO->getSyncScopeID(); |
| Ordering = |
| isStrongerThan(Ordering, OpOrdering) ? |
| Ordering : MMO->getOrdering(); |
| assert(MMO->getFailureOrdering() != AtomicOrdering::Release && |
| MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease); |
| FailureOrdering = |
| isStrongerThan(FailureOrdering, MMO->getFailureOrdering()) ? |
| FailureOrdering : MMO->getFailureOrdering(); |
| } |
| } |
| |
| SIAtomicScope Scope = SIAtomicScope::NONE; |
| SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; |
| bool IsCrossAddressSpaceOrdering = false; |
| if (Ordering != AtomicOrdering::NotAtomic) { |
| auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace); |
| if (!ScopeOrNone) { |
| reportUnsupported(MI, "Unsupported atomic synchronization scope"); |
| return None; |
| } |
| std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = |
| ScopeOrNone.getValue(); |
| if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || |
| ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { |
| reportUnsupported(MI, "Unsupported atomic address space"); |
| return None; |
| } |
| } |
| return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, InstrAddrSpace, |
| IsCrossAddressSpaceOrdering, FailureOrdering, IsNonTemporal); |
| } |
| |
| Optional<SIMemOpInfo> SIMemOpAccess::getLoadInfo( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); |
| |
| if (!(MI->mayLoad() && !MI->mayStore())) |
| return None; |
| |
| // Be conservative if there are no memory operands. |
| if (MI->getNumMemOperands() == 0) |
| return SIMemOpInfo(); |
| |
| return constructFromMIWithMMO(MI); |
| } |
| |
| Optional<SIMemOpInfo> SIMemOpAccess::getStoreInfo( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); |
| |
| if (!(!MI->mayLoad() && MI->mayStore())) |
| return None; |
| |
| // Be conservative if there are no memory operands. |
| if (MI->getNumMemOperands() == 0) |
| return SIMemOpInfo(); |
| |
| return constructFromMIWithMMO(MI); |
| } |
| |
| Optional<SIMemOpInfo> SIMemOpAccess::getAtomicFenceInfo( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); |
| |
| if (MI->getOpcode() != AMDGPU::ATOMIC_FENCE) |
| return None; |
| |
| AtomicOrdering Ordering = |
| static_cast<AtomicOrdering>(MI->getOperand(0).getImm()); |
| |
| SyncScope::ID SSID = static_cast<SyncScope::ID>(MI->getOperand(1).getImm()); |
| auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC); |
| if (!ScopeOrNone) { |
| reportUnsupported(MI, "Unsupported atomic synchronization scope"); |
| return None; |
| } |
| |
| SIAtomicScope Scope = SIAtomicScope::NONE; |
| SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE; |
| bool IsCrossAddressSpaceOrdering = false; |
| std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) = |
| ScopeOrNone.getValue(); |
| |
| if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) || |
| ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) { |
| reportUnsupported(MI, "Unsupported atomic address space"); |
| return None; |
| } |
| |
| return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, SIAtomicAddrSpace::ATOMIC, |
| IsCrossAddressSpaceOrdering); |
| } |
| |
| Optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic); |
| |
| if (!(MI->mayLoad() && MI->mayStore())) |
| return None; |
| |
| // Be conservative if there are no memory operands. |
| if (MI->getNumMemOperands() == 0) |
| return SIMemOpInfo(); |
| |
| return constructFromMIWithMMO(MI); |
| } |
| |
| SICacheControl::SICacheControl(const GCNSubtarget &ST) { |
| TII = ST.getInstrInfo(); |
| IV = IsaInfo::getIsaVersion(ST.getFeatureBits()); |
| } |
| |
| /* static */ |
| std::unique_ptr<SICacheControl> SICacheControl::create(const GCNSubtarget &ST) { |
| GCNSubtarget::Generation Generation = ST.getGeneration(); |
| if (Generation <= AMDGPUSubtarget::SOUTHERN_ISLANDS) |
| return make_unique<SIGfx6CacheControl>(ST); |
| return make_unique<SIGfx7CacheControl>(ST); |
| } |
| |
| bool SIGfx6CacheControl::enableLoadCacheBypass( |
| const MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace) const { |
| assert(MI->mayLoad() && !MI->mayStore()); |
| bool Changed = false; |
| |
| if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { |
| /// TODO: Do not set glc for rmw atomic operations as they |
| /// implicitly bypass the L1 cache. |
| |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| Changed |= enableGLCBit(MI); |
| break; |
| case SIAtomicScope::WORKGROUP: |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // No cache to bypass. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| /// The scratch address space does not need the global memory caches |
| /// to be bypassed as all memory operations by the same thread are |
| /// sequentially consistent, and no other thread can access scratch |
| /// memory. |
| |
| /// Other address spaces do not hava a cache. |
| |
| return Changed; |
| } |
| |
| bool SIGfx6CacheControl::enableNonTemporal( |
| const MachineBasicBlock::iterator &MI) const { |
| assert(MI->mayLoad() ^ MI->mayStore()); |
| bool Changed = false; |
| |
| /// TODO: Do not enableGLCBit if rmw atomic. |
| Changed |= enableGLCBit(MI); |
| Changed |= enableSLCBit(MI); |
| |
| return Changed; |
| } |
| |
| bool SIGfx6CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| Position Pos) const { |
| bool Changed = false; |
| |
| MachineBasicBlock &MBB = *MI->getParent(); |
| DebugLoc DL = MI->getDebugLoc(); |
| |
| if (Pos == Position::AFTER) |
| ++MI; |
| |
| if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1)); |
| Changed = true; |
| break; |
| case SIAtomicScope::WORKGROUP: |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // No cache to invalidate. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| /// The scratch address space does not need the global memory cache |
| /// to be flushed as all memory operations by the same thread are |
| /// sequentially consistent, and no other thread can access scratch |
| /// memory. |
| |
| /// Other address spaces do not hava a cache. |
| |
| if (Pos == Position::AFTER) |
| --MI; |
| |
| return Changed; |
| } |
| |
| bool SIGfx6CacheControl::insertWait(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| SIMemOp Op, |
| bool IsCrossAddrSpaceOrdering, |
| Position Pos) const { |
| bool Changed = false; |
| |
| MachineBasicBlock &MBB = *MI->getParent(); |
| DebugLoc DL = MI->getDebugLoc(); |
| |
| if (Pos == Position::AFTER) |
| ++MI; |
| |
| bool VMCnt = false; |
| bool LGKMCnt = false; |
| bool EXPCnt = false; |
| |
| if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| VMCnt = true; |
| break; |
| case SIAtomicScope::WORKGROUP: |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // The L1 cache keeps all memory operations in order for |
| // wavefronts in the same work-group. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) { |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| case SIAtomicScope::WORKGROUP: |
| // If no cross address space ordering then an LDS waitcnt is not |
| // needed as LDS operations for all waves are executed in a |
| // total global ordering as observed by all waves. Required if |
| // also synchronizing with global/GDS memory as LDS operations |
| // could be reordered with respect to later global/GDS memory |
| // operations of the same wave. |
| LGKMCnt = IsCrossAddrSpaceOrdering; |
| break; |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // The LDS keeps all memory operations in order for |
| // the same wavesfront. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) { |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| // If no cross address space ordering then an GDS waitcnt is not |
| // needed as GDS operations for all waves are executed in a |
| // total global ordering as observed by all waves. Required if |
| // also synchronizing with global/LDS memory as GDS operations |
| // could be reordered with respect to later global/LDS memory |
| // operations of the same wave. |
| EXPCnt = IsCrossAddrSpaceOrdering; |
| break; |
| case SIAtomicScope::WORKGROUP: |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // The GDS keeps all memory operations in order for |
| // the same work-group. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| if (VMCnt || LGKMCnt || EXPCnt) { |
| unsigned WaitCntImmediate = |
| AMDGPU::encodeWaitcnt(IV, |
| VMCnt ? 0 : getVmcntBitMask(IV), |
| EXPCnt ? 0 : getExpcntBitMask(IV), |
| LGKMCnt ? 0 : getLgkmcntBitMask(IV)); |
| BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate); |
| Changed = true; |
| } |
| |
| if (Pos == Position::AFTER) |
| --MI; |
| |
| return Changed; |
| } |
| |
| bool SIGfx7CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI, |
| SIAtomicScope Scope, |
| SIAtomicAddrSpace AddrSpace, |
| Position Pos) const { |
| bool Changed = false; |
| |
| MachineBasicBlock &MBB = *MI->getParent(); |
| DebugLoc DL = MI->getDebugLoc(); |
| |
| if (Pos == Position::AFTER) |
| ++MI; |
| |
| if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) { |
| switch (Scope) { |
| case SIAtomicScope::SYSTEM: |
| case SIAtomicScope::AGENT: |
| BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1_VOL)); |
| Changed = true; |
| break; |
| case SIAtomicScope::WORKGROUP: |
| case SIAtomicScope::WAVEFRONT: |
| case SIAtomicScope::SINGLETHREAD: |
| // No cache to invalidate. |
| break; |
| default: |
| llvm_unreachable("Unsupported synchronization scope"); |
| } |
| } |
| |
| /// The scratch address space does not need the global memory cache |
| /// to be flushed as all memory operations by the same thread are |
| /// sequentially consistent, and no other thread can access scratch |
| /// memory. |
| |
| /// Other address spaces do not hava a cache. |
| |
| if (Pos == Position::AFTER) |
| --MI; |
| |
| return Changed; |
| } |
| |
| bool SIMemoryLegalizer::removeAtomicPseudoMIs() { |
| if (AtomicPseudoMIs.empty()) |
| return false; |
| |
| for (auto &MI : AtomicPseudoMIs) |
| MI->eraseFromParent(); |
| |
| AtomicPseudoMIs.clear(); |
| return true; |
| } |
| |
| bool SIMemoryLegalizer::expandLoad(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI) { |
| assert(MI->mayLoad() && !MI->mayStore()); |
| |
| bool Changed = false; |
| |
| if (MOI.isAtomic()) { |
| if (MOI.getOrdering() == AtomicOrdering::Monotonic || |
| MOI.getOrdering() == AtomicOrdering::Acquire || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { |
| Changed |= CC->enableLoadCacheBypass(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace()); |
| } |
| |
| if (MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| SIMemOp::LOAD | SIMemOp::STORE, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::BEFORE); |
| |
| if (MOI.getOrdering() == AtomicOrdering::Acquire || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) { |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getInstrAddrSpace(), |
| SIMemOp::LOAD, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::AFTER); |
| Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| Position::AFTER); |
| } |
| |
| return Changed; |
| } |
| |
| // Atomic instructions do not have the nontemporal attribute. |
| if (MOI.isNonTemporal()) { |
| Changed |= CC->enableNonTemporal(MI); |
| return Changed; |
| } |
| |
| return Changed; |
| } |
| |
| bool SIMemoryLegalizer::expandStore(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI) { |
| assert(!MI->mayLoad() && MI->mayStore()); |
| |
| bool Changed = false; |
| |
| if (MOI.isAtomic()) { |
| if (MOI.getOrdering() == AtomicOrdering::Release || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| SIMemOp::LOAD | SIMemOp::STORE, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::BEFORE); |
| |
| return Changed; |
| } |
| |
| // Atomic instructions do not have the nontemporal attribute. |
| if (MOI.isNonTemporal()) { |
| Changed |= CC->enableNonTemporal(MI); |
| return Changed; |
| } |
| |
| return Changed; |
| } |
| |
| bool SIMemoryLegalizer::expandAtomicFence(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI) { |
| assert(MI->getOpcode() == AMDGPU::ATOMIC_FENCE); |
| |
| AtomicPseudoMIs.push_back(MI); |
| bool Changed = false; |
| |
| if (MOI.isAtomic()) { |
| if (MOI.getOrdering() == AtomicOrdering::Acquire || |
| MOI.getOrdering() == AtomicOrdering::Release || |
| MOI.getOrdering() == AtomicOrdering::AcquireRelease || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) |
| /// TODO: This relies on a barrier always generating a waitcnt |
| /// for LDS to ensure it is not reordered with the completion of |
| /// the proceeding LDS operations. If barrier had a memory |
| /// ordering and memory scope, then library does not need to |
| /// generate a fence. Could add support in this file for |
| /// barrier. SIInsertWaitcnt.cpp could then stop unconditionally |
| /// adding waitcnt before a S_BARRIER. |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| SIMemOp::LOAD | SIMemOp::STORE, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::BEFORE); |
| |
| if (MOI.getOrdering() == AtomicOrdering::Acquire || |
| MOI.getOrdering() == AtomicOrdering::AcquireRelease || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) |
| Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| Position::BEFORE); |
| |
| return Changed; |
| } |
| |
| return Changed; |
| } |
| |
| bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI, |
| MachineBasicBlock::iterator &MI) { |
| assert(MI->mayLoad() && MI->mayStore()); |
| |
| bool Changed = false; |
| |
| if (MOI.isAtomic()) { |
| if (MOI.getOrdering() == AtomicOrdering::Release || |
| MOI.getOrdering() == AtomicOrdering::AcquireRelease || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || |
| MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| SIMemOp::LOAD | SIMemOp::STORE, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::BEFORE); |
| |
| if (MOI.getOrdering() == AtomicOrdering::Acquire || |
| MOI.getOrdering() == AtomicOrdering::AcquireRelease || |
| MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent || |
| MOI.getFailureOrdering() == AtomicOrdering::Acquire || |
| MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) { |
| Changed |= CC->insertWait(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| isAtomicRet(*MI) ? SIMemOp::LOAD : |
| SIMemOp::STORE, |
| MOI.getIsCrossAddressSpaceOrdering(), |
| Position::AFTER); |
| Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(), |
| MOI.getOrderingAddrSpace(), |
| Position::AFTER); |
| } |
| |
| return Changed; |
| } |
| |
| return Changed; |
| } |
| |
| bool SIMemoryLegalizer::runOnMachineFunction(MachineFunction &MF) { |
| bool Changed = false; |
| |
| SIMemOpAccess MOA(MF); |
| CC = SICacheControl::create(MF.getSubtarget<GCNSubtarget>()); |
| |
| for (auto &MBB : MF) { |
| for (auto MI = MBB.begin(); MI != MBB.end(); ++MI) { |
| if (!(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic)) |
| continue; |
| |
| if (const auto &MOI = MOA.getLoadInfo(MI)) |
| Changed |= expandLoad(MOI.getValue(), MI); |
| else if (const auto &MOI = MOA.getStoreInfo(MI)) |
| Changed |= expandStore(MOI.getValue(), MI); |
| else if (const auto &MOI = MOA.getAtomicFenceInfo(MI)) |
| Changed |= expandAtomicFence(MOI.getValue(), MI); |
| else if (const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(MI)) |
| Changed |= expandAtomicCmpxchgOrRmw(MOI.getValue(), MI); |
| } |
| } |
| |
| Changed |= removeAtomicPseudoMIs(); |
| return Changed; |
| } |
| |
| INITIALIZE_PASS(SIMemoryLegalizer, DEBUG_TYPE, PASS_NAME, false, false) |
| |
| char SIMemoryLegalizer::ID = 0; |
| char &llvm::SIMemoryLegalizerID = SIMemoryLegalizer::ID; |
| |
| FunctionPass *llvm::createSIMemoryLegalizerPass() { |
| return new SIMemoryLegalizer(); |
| } |