forked from OSchip/llvm-project
				
			
		
			
				
	
	
		
			1042 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			1042 lines
		
	
	
		
			36 KiB
		
	
	
	
		
			C++
		
	
	
	
//===- SIMemoryLegalizer.cpp ----------------------------------------------===//
 | 
						|
//
 | 
						|
//                     The LLVM Compiler Infrastructure
 | 
						|
//
 | 
						|
// This file is distributed under the University of Illinois Open Source
 | 
						|
// License. See LICENSE.TXT for details.
 | 
						|
//
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
//
 | 
						|
/// \file
 | 
						|
/// Memory legalizer - implements memory model. More information can be
 | 
						|
/// found here:
 | 
						|
///   http://llvm.org/docs/AMDGPUUsage.html#memory-model
 | 
						|
//
 | 
						|
//===----------------------------------------------------------------------===//
 | 
						|
 | 
						|
#include "AMDGPU.h"
 | 
						|
#include "AMDGPUMachineModuleInfo.h"
 | 
						|
#include "AMDGPUSubtarget.h"
 | 
						|
#include "SIDefines.h"
 | 
						|
#include "SIInstrInfo.h"
 | 
						|
#include "MCTargetDesc/AMDGPUMCTargetDesc.h"
 | 
						|
#include "Utils/AMDGPUBaseInfo.h"
 | 
						|
#include "llvm/ADT/BitmaskEnum.h"
 | 
						|
#include "llvm/ADT/None.h"
 | 
						|
#include "llvm/ADT/Optional.h"
 | 
						|
#include "llvm/CodeGen/MachineBasicBlock.h"
 | 
						|
#include "llvm/CodeGen/MachineFunction.h"
 | 
						|
#include "llvm/CodeGen/MachineFunctionPass.h"
 | 
						|
#include "llvm/CodeGen/MachineInstrBuilder.h"
 | 
						|
#include "llvm/CodeGen/MachineMemOperand.h"
 | 
						|
#include "llvm/CodeGen/MachineModuleInfo.h"
 | 
						|
#include "llvm/CodeGen/MachineOperand.h"
 | 
						|
#include "llvm/IR/DebugLoc.h"
 | 
						|
#include "llvm/IR/DiagnosticInfo.h"
 | 
						|
#include "llvm/IR/Function.h"
 | 
						|
#include "llvm/IR/LLVMContext.h"
 | 
						|
#include "llvm/MC/MCInstrDesc.h"
 | 
						|
#include "llvm/Pass.h"
 | 
						|
#include "llvm/Support/AtomicOrdering.h"
 | 
						|
#include "llvm/Support/MathExtras.h"
 | 
						|
#include <cassert>
 | 
						|
#include <list>
 | 
						|
 | 
						|
using namespace llvm;
 | 
						|
using namespace llvm::AMDGPU;
 | 
						|
 | 
						|
#define DEBUG_TYPE "si-memory-legalizer"
 | 
						|
#define PASS_NAME "SI Memory Legalizer"
 | 
						|
 | 
						|
namespace {
 | 
						|
 | 
						|
LLVM_ENABLE_BITMASK_ENUMS_IN_NAMESPACE();
 | 
						|
 | 
						|
/// Memory operation flags. Can be ORed together.
 | 
						|
enum class SIMemOp {
 | 
						|
  NONE = 0u,
 | 
						|
  LOAD = 1u << 0,
 | 
						|
  STORE = 1u << 1,
 | 
						|
  LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ STORE)
 | 
						|
};
 | 
						|
 | 
						|
/// Position to insert a new instruction relative to an existing
 | 
						|
/// instruction.
 | 
						|
enum class Position {
 | 
						|
  BEFORE,
 | 
						|
  AFTER
 | 
						|
};
 | 
						|
 | 
						|
/// The atomic synchronization scopes supported by the AMDGPU target.
 | 
						|
enum class SIAtomicScope {
 | 
						|
  NONE,
 | 
						|
  SINGLETHREAD,
 | 
						|
  WAVEFRONT,
 | 
						|
  WORKGROUP,
 | 
						|
  AGENT,
 | 
						|
  SYSTEM
 | 
						|
};
 | 
						|
 | 
						|
/// The distinct address spaces supported by the AMDGPU target for
 | 
						|
/// atomic memory operation. Can be ORed toether.
 | 
						|
enum class SIAtomicAddrSpace {
 | 
						|
  NONE = 0u,
 | 
						|
  GLOBAL = 1u << 0,
 | 
						|
  LDS = 1u << 1,
 | 
						|
  SCRATCH = 1u << 2,
 | 
						|
  GDS = 1u << 3,
 | 
						|
  OTHER = 1u << 4,
 | 
						|
 | 
						|
  /// The address spaces that can be accessed by a FLAT instruction.
 | 
						|
  FLAT = GLOBAL | LDS | SCRATCH,
 | 
						|
 | 
						|
  /// The address spaces that support atomic instructions.
 | 
						|
  ATOMIC = GLOBAL | LDS | SCRATCH | GDS,
 | 
						|
 | 
						|
  /// All address spaces.
 | 
						|
  ALL = GLOBAL | LDS | SCRATCH | GDS | OTHER,
 | 
						|
 | 
						|
  LLVM_MARK_AS_BITMASK_ENUM(/* LargestFlag = */ ALL)
 | 
						|
};
 | 
						|
 | 
						|
/// Sets named bit \p BitName to "true" if present in instruction \p MI.
 | 
						|
/// \returns Returns true if \p MI is modified, false otherwise.
 | 
						|
template <uint16_t BitName>
 | 
						|
bool enableNamedBit(const MachineBasicBlock::iterator &MI) {
 | 
						|
  int BitIdx = AMDGPU::getNamedOperandIdx(MI->getOpcode(), BitName);
 | 
						|
  if (BitIdx == -1)
 | 
						|
    return false;
 | 
						|
 | 
						|
  MachineOperand &Bit = MI->getOperand(BitIdx);
 | 
						|
  if (Bit.getImm() != 0)
 | 
						|
    return false;
 | 
						|
 | 
						|
  Bit.setImm(1);
 | 
						|
  return true;
 | 
						|
}
 | 
						|
 | 
						|
class SIMemOpInfo final {
 | 
						|
private:
 | 
						|
 | 
						|
  friend class SIMemOpAccess;
 | 
						|
 | 
						|
  AtomicOrdering Ordering = AtomicOrdering::NotAtomic;
 | 
						|
  AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic;
 | 
						|
  SIAtomicScope Scope = SIAtomicScope::SYSTEM;
 | 
						|
  SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
 | 
						|
  SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
 | 
						|
  bool IsCrossAddressSpaceOrdering = false;
 | 
						|
  bool IsNonTemporal = false;
 | 
						|
 | 
						|
  SIMemOpInfo(AtomicOrdering Ordering = AtomicOrdering::SequentiallyConsistent,
 | 
						|
              SIAtomicScope Scope = SIAtomicScope::SYSTEM,
 | 
						|
              SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::ATOMIC,
 | 
						|
              SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::ALL,
 | 
						|
              bool IsCrossAddressSpaceOrdering = true,
 | 
						|
              AtomicOrdering FailureOrdering =
 | 
						|
                AtomicOrdering::SequentiallyConsistent,
 | 
						|
              bool IsNonTemporal = false)
 | 
						|
    : Ordering(Ordering), FailureOrdering(FailureOrdering),
 | 
						|
      Scope(Scope), OrderingAddrSpace(OrderingAddrSpace),
 | 
						|
      InstrAddrSpace(InstrAddrSpace),
 | 
						|
      IsCrossAddressSpaceOrdering(IsCrossAddressSpaceOrdering),
 | 
						|
      IsNonTemporal(IsNonTemporal) {
 | 
						|
    // There is also no cross address space ordering if the ordering
 | 
						|
    // address space is the same as the instruction address space and
 | 
						|
    // only contains a single address space.
 | 
						|
    if ((OrderingAddrSpace == InstrAddrSpace) &&
 | 
						|
        isPowerOf2_32(uint32_t(InstrAddrSpace)))
 | 
						|
      IsCrossAddressSpaceOrdering = false;
 | 
						|
  }
 | 
						|
 | 
						|
public:
 | 
						|
  /// \returns Atomic synchronization scope of the machine instruction used to
 | 
						|
  /// create this SIMemOpInfo.
 | 
						|
  SIAtomicScope getScope() const {
 | 
						|
    return Scope;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns Ordering constraint of the machine instruction used to
 | 
						|
  /// create this SIMemOpInfo.
 | 
						|
  AtomicOrdering getOrdering() const {
 | 
						|
    return Ordering;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns Failure ordering constraint of the machine instruction used to
 | 
						|
  /// create this SIMemOpInfo.
 | 
						|
  AtomicOrdering getFailureOrdering() const {
 | 
						|
    return FailureOrdering;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns The address spaces be accessed by the machine
 | 
						|
  /// instruction used to create this SiMemOpInfo.
 | 
						|
  SIAtomicAddrSpace getInstrAddrSpace() const {
 | 
						|
    return InstrAddrSpace;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns The address spaces that must be ordered by the machine
 | 
						|
  /// instruction used to create this SiMemOpInfo.
 | 
						|
  SIAtomicAddrSpace getOrderingAddrSpace() const {
 | 
						|
    return OrderingAddrSpace;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns Return true iff memory ordering of operations on
 | 
						|
  /// different address spaces is required.
 | 
						|
  bool getIsCrossAddressSpaceOrdering() const {
 | 
						|
    return IsCrossAddressSpaceOrdering;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns True if memory access of the machine instruction used to
 | 
						|
  /// create this SIMemOpInfo is non-temporal, false otherwise.
 | 
						|
  bool isNonTemporal() const {
 | 
						|
    return IsNonTemporal;
 | 
						|
  }
 | 
						|
 | 
						|
  /// \returns True if ordering constraint of the machine instruction used to
 | 
						|
  /// create this SIMemOpInfo is unordered or higher, false otherwise.
 | 
						|
  bool isAtomic() const {
 | 
						|
    return Ordering != AtomicOrdering::NotAtomic;
 | 
						|
  }
 | 
						|
 | 
						|
};
 | 
						|
 | 
						|
class SIMemOpAccess final {
 | 
						|
private:
 | 
						|
  AMDGPUMachineModuleInfo *MMI = nullptr;
 | 
						|
 | 
						|
  /// Reports unsupported message \p Msg for \p MI to LLVM context.
 | 
						|
  void reportUnsupported(const MachineBasicBlock::iterator &MI,
 | 
						|
                         const char *Msg) const;
 | 
						|
 | 
						|
  /// Inspects the target synchonization scope \p SSID and determines
 | 
						|
  /// the SI atomic scope it corresponds to, the address spaces it
 | 
						|
  /// covers, and whether the memory ordering applies between address
 | 
						|
  /// spaces.
 | 
						|
  Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
 | 
						|
  toSIAtomicScope(SyncScope::ID SSID, SIAtomicAddrSpace InstrScope) const;
 | 
						|
 | 
						|
  /// \return Return a bit set of the address spaces accessed by \p AS.
 | 
						|
  SIAtomicAddrSpace toSIAtomicAddrSpace(unsigned AS) const;
 | 
						|
 | 
						|
  /// \returns Info constructed from \p MI, which has at least machine memory
 | 
						|
  /// operand.
 | 
						|
  Optional<SIMemOpInfo> constructFromMIWithMMO(
 | 
						|
      const MachineBasicBlock::iterator &MI) const;
 | 
						|
 | 
						|
public:
 | 
						|
  /// Construct class to support accessing the machine memory operands
 | 
						|
  /// of instructions in the machine function \p MF.
 | 
						|
  SIMemOpAccess(MachineFunction &MF);
 | 
						|
 | 
						|
  /// \returns Load info if \p MI is a load operation, "None" otherwise.
 | 
						|
  Optional<SIMemOpInfo> getLoadInfo(
 | 
						|
      const MachineBasicBlock::iterator &MI) const;
 | 
						|
 | 
						|
  /// \returns Store info if \p MI is a store operation, "None" otherwise.
 | 
						|
  Optional<SIMemOpInfo> getStoreInfo(
 | 
						|
      const MachineBasicBlock::iterator &MI) const;
 | 
						|
 | 
						|
  /// \returns Atomic fence info if \p MI is an atomic fence operation,
 | 
						|
  /// "None" otherwise.
 | 
						|
  Optional<SIMemOpInfo> getAtomicFenceInfo(
 | 
						|
      const MachineBasicBlock::iterator &MI) const;
 | 
						|
 | 
						|
  /// \returns Atomic cmpxchg/rmw info if \p MI is an atomic cmpxchg or
 | 
						|
  /// rmw operation, "None" otherwise.
 | 
						|
  Optional<SIMemOpInfo> getAtomicCmpxchgOrRmwInfo(
 | 
						|
      const MachineBasicBlock::iterator &MI) const;
 | 
						|
};
 | 
						|
 | 
						|
class SICacheControl {
 | 
						|
protected:
 | 
						|
 | 
						|
  /// Instruction info.
 | 
						|
  const SIInstrInfo *TII = nullptr;
 | 
						|
 | 
						|
  IsaVersion IV;
 | 
						|
 | 
						|
  SICacheControl(const GCNSubtarget &ST);
 | 
						|
 | 
						|
public:
 | 
						|
 | 
						|
  /// Create a cache control for the subtarget \p ST.
 | 
						|
  static std::unique_ptr<SICacheControl> create(const GCNSubtarget &ST);
 | 
						|
 | 
						|
  /// Update \p MI memory load instruction to bypass any caches up to
 | 
						|
  /// the \p Scope memory scope for address spaces \p
 | 
						|
  /// AddrSpace. Return true iff the instruction was modified.
 | 
						|
  virtual bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI,
 | 
						|
                                     SIAtomicScope Scope,
 | 
						|
                                     SIAtomicAddrSpace AddrSpace) const = 0;
 | 
						|
 | 
						|
  /// Update \p MI memory instruction to indicate it is
 | 
						|
  /// nontemporal. Return true iff the instruction was modified.
 | 
						|
  virtual bool enableNonTemporal(const MachineBasicBlock::iterator &MI)
 | 
						|
    const = 0;
 | 
						|
 | 
						|
  /// Inserts any necessary instructions at position \p Pos relative
 | 
						|
  /// to instruction \p MI to ensure any caches associated with
 | 
						|
  /// address spaces \p AddrSpace for memory scopes up to memory scope
 | 
						|
  /// \p Scope are invalidated. Returns true iff any instructions
 | 
						|
  /// inserted.
 | 
						|
  virtual bool insertCacheInvalidate(MachineBasicBlock::iterator &MI,
 | 
						|
                                     SIAtomicScope Scope,
 | 
						|
                                     SIAtomicAddrSpace AddrSpace,
 | 
						|
                                     Position Pos) const = 0;
 | 
						|
 | 
						|
  /// Inserts any necessary instructions at position \p Pos relative
 | 
						|
  /// to instruction \p MI to ensure memory instructions of kind \p Op
 | 
						|
  /// associated with address spaces \p AddrSpace have completed as
 | 
						|
  /// observed by other memory instructions executing in memory scope
 | 
						|
  /// \p Scope. \p IsCrossAddrSpaceOrdering indicates if the memory
 | 
						|
  /// ordering is between address spaces. Returns true iff any
 | 
						|
  /// instructions inserted.
 | 
						|
  virtual bool insertWait(MachineBasicBlock::iterator &MI,
 | 
						|
                          SIAtomicScope Scope,
 | 
						|
                          SIAtomicAddrSpace AddrSpace,
 | 
						|
                          SIMemOp Op,
 | 
						|
                          bool IsCrossAddrSpaceOrdering,
 | 
						|
                          Position Pos) const = 0;
 | 
						|
 | 
						|
  /// Virtual destructor to allow derivations to be deleted.
 | 
						|
  virtual ~SICacheControl() = default;
 | 
						|
 | 
						|
};
 | 
						|
 | 
						|
class SIGfx6CacheControl : public SICacheControl {
 | 
						|
protected:
 | 
						|
 | 
						|
  /// Sets GLC bit to "true" if present in \p MI. Returns true if \p MI
 | 
						|
  /// is modified, false otherwise.
 | 
						|
  bool enableGLCBit(const MachineBasicBlock::iterator &MI) const {
 | 
						|
    return enableNamedBit<AMDGPU::OpName::glc>(MI);
 | 
						|
  }
 | 
						|
 | 
						|
  /// Sets SLC bit to "true" if present in \p MI. Returns true if \p MI
 | 
						|
  /// is modified, false otherwise.
 | 
						|
  bool enableSLCBit(const MachineBasicBlock::iterator &MI) const {
 | 
						|
    return enableNamedBit<AMDGPU::OpName::slc>(MI);
 | 
						|
  }
 | 
						|
 | 
						|
public:
 | 
						|
 | 
						|
  SIGfx6CacheControl(const GCNSubtarget &ST) : SICacheControl(ST) {};
 | 
						|
 | 
						|
  bool enableLoadCacheBypass(const MachineBasicBlock::iterator &MI,
 | 
						|
                             SIAtomicScope Scope,
 | 
						|
                             SIAtomicAddrSpace AddrSpace) const override;
 | 
						|
 | 
						|
  bool enableNonTemporal(const MachineBasicBlock::iterator &MI) const override;
 | 
						|
 | 
						|
  bool insertCacheInvalidate(MachineBasicBlock::iterator &MI,
 | 
						|
                             SIAtomicScope Scope,
 | 
						|
                             SIAtomicAddrSpace AddrSpace,
 | 
						|
                             Position Pos) const override;
 | 
						|
 | 
						|
  bool insertWait(MachineBasicBlock::iterator &MI,
 | 
						|
                  SIAtomicScope Scope,
 | 
						|
                  SIAtomicAddrSpace AddrSpace,
 | 
						|
                  SIMemOp Op,
 | 
						|
                  bool IsCrossAddrSpaceOrdering,
 | 
						|
                  Position Pos) const override;
 | 
						|
};
 | 
						|
 | 
						|
class SIGfx7CacheControl : public SIGfx6CacheControl {
 | 
						|
public:
 | 
						|
 | 
						|
  SIGfx7CacheControl(const GCNSubtarget &ST) : SIGfx6CacheControl(ST) {};
 | 
						|
 | 
						|
  bool insertCacheInvalidate(MachineBasicBlock::iterator &MI,
 | 
						|
                             SIAtomicScope Scope,
 | 
						|
                             SIAtomicAddrSpace AddrSpace,
 | 
						|
                             Position Pos) const override;
 | 
						|
 | 
						|
};
 | 
						|
 | 
						|
class SIMemoryLegalizer final : public MachineFunctionPass {
 | 
						|
private:
 | 
						|
 | 
						|
  /// Cache Control.
 | 
						|
  std::unique_ptr<SICacheControl> CC = nullptr;
 | 
						|
 | 
						|
  /// List of atomic pseudo instructions.
 | 
						|
  std::list<MachineBasicBlock::iterator> AtomicPseudoMIs;
 | 
						|
 | 
						|
  /// Return true iff instruction \p MI is a atomic instruction that
 | 
						|
  /// returns a result.
 | 
						|
  bool isAtomicRet(const MachineInstr &MI) const {
 | 
						|
    return AMDGPU::getAtomicNoRetOp(MI.getOpcode()) != -1;
 | 
						|
  }
 | 
						|
 | 
						|
  /// Removes all processed atomic pseudo instructions from the current
 | 
						|
  /// function. Returns true if current function is modified, false otherwise.
 | 
						|
  bool removeAtomicPseudoMIs();
 | 
						|
 | 
						|
  /// Expands load operation \p MI. Returns true if instructions are
 | 
						|
  /// added/deleted or \p MI is modified, false otherwise.
 | 
						|
  bool expandLoad(const SIMemOpInfo &MOI,
 | 
						|
                  MachineBasicBlock::iterator &MI);
 | 
						|
  /// Expands store operation \p MI. Returns true if instructions are
 | 
						|
  /// added/deleted or \p MI is modified, false otherwise.
 | 
						|
  bool expandStore(const SIMemOpInfo &MOI,
 | 
						|
                   MachineBasicBlock::iterator &MI);
 | 
						|
  /// Expands atomic fence operation \p MI. Returns true if
 | 
						|
  /// instructions are added/deleted or \p MI is modified, false otherwise.
 | 
						|
  bool expandAtomicFence(const SIMemOpInfo &MOI,
 | 
						|
                         MachineBasicBlock::iterator &MI);
 | 
						|
  /// Expands atomic cmpxchg or rmw operation \p MI. Returns true if
 | 
						|
  /// instructions are added/deleted or \p MI is modified, false otherwise.
 | 
						|
  bool expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI,
 | 
						|
                                MachineBasicBlock::iterator &MI);
 | 
						|
 | 
						|
public:
 | 
						|
  static char ID;
 | 
						|
 | 
						|
  SIMemoryLegalizer() : MachineFunctionPass(ID) {}
 | 
						|
 | 
						|
  void getAnalysisUsage(AnalysisUsage &AU) const override {
 | 
						|
    AU.setPreservesCFG();
 | 
						|
    MachineFunctionPass::getAnalysisUsage(AU);
 | 
						|
  }
 | 
						|
 | 
						|
  StringRef getPassName() const override {
 | 
						|
    return PASS_NAME;
 | 
						|
  }
 | 
						|
 | 
						|
  bool runOnMachineFunction(MachineFunction &MF) override;
 | 
						|
};
 | 
						|
 | 
						|
} // end namespace anonymous
 | 
						|
 | 
						|
void SIMemOpAccess::reportUnsupported(const MachineBasicBlock::iterator &MI,
 | 
						|
                                      const char *Msg) const {
 | 
						|
  const Function &Func = MI->getParent()->getParent()->getFunction();
 | 
						|
  DiagnosticInfoUnsupported Diag(Func, Msg, MI->getDebugLoc());
 | 
						|
  Func.getContext().diagnose(Diag);
 | 
						|
}
 | 
						|
 | 
						|
Optional<std::tuple<SIAtomicScope, SIAtomicAddrSpace, bool>>
 | 
						|
SIMemOpAccess::toSIAtomicScope(SyncScope::ID SSID,
 | 
						|
                               SIAtomicAddrSpace InstrScope) const {
 | 
						|
  /// TODO: For now assume OpenCL memory model which treats each
 | 
						|
  /// address space as having a separate happens-before relation, and
 | 
						|
  /// so an instruction only has ordering with respect to the address
 | 
						|
  /// space it accesses, and if it accesses multiple address spaces it
 | 
						|
  /// does not require ordering of operations in different address
 | 
						|
  /// spaces.
 | 
						|
 if (SSID == SyncScope::System)
 | 
						|
    return std::make_tuple(SIAtomicScope::SYSTEM,
 | 
						|
                           SIAtomicAddrSpace::ATOMIC & InstrScope,
 | 
						|
                           false);
 | 
						|
  if (SSID == MMI->getAgentSSID())
 | 
						|
    return std::make_tuple(SIAtomicScope::AGENT,
 | 
						|
                           SIAtomicAddrSpace::ATOMIC & InstrScope,
 | 
						|
                           false);
 | 
						|
  if (SSID == MMI->getWorkgroupSSID())
 | 
						|
    return std::make_tuple(SIAtomicScope::WORKGROUP,
 | 
						|
                           SIAtomicAddrSpace::ATOMIC & InstrScope,
 | 
						|
                           false);
 | 
						|
  if (SSID == MMI->getWavefrontSSID())
 | 
						|
    return std::make_tuple(SIAtomicScope::WAVEFRONT,
 | 
						|
                           SIAtomicAddrSpace::ATOMIC & InstrScope,
 | 
						|
                           false);
 | 
						|
  if (SSID == SyncScope::SingleThread)
 | 
						|
    return std::make_tuple(SIAtomicScope::SINGLETHREAD,
 | 
						|
                           SIAtomicAddrSpace::ATOMIC & InstrScope,
 | 
						|
                           false);
 | 
						|
  /// TODO: To support HSA Memory Model need to add additional memory
 | 
						|
  /// scopes that specify that do require cross address space
 | 
						|
  /// ordering.
 | 
						|
  return None;
 | 
						|
}
 | 
						|
 | 
						|
SIAtomicAddrSpace SIMemOpAccess::toSIAtomicAddrSpace(unsigned AS) const {
 | 
						|
  if (AS == AMDGPUAS::FLAT_ADDRESS)
 | 
						|
    return SIAtomicAddrSpace::FLAT;
 | 
						|
  if (AS == AMDGPUAS::GLOBAL_ADDRESS)
 | 
						|
    return SIAtomicAddrSpace::GLOBAL;
 | 
						|
  if (AS == AMDGPUAS::LOCAL_ADDRESS)
 | 
						|
    return SIAtomicAddrSpace::LDS;
 | 
						|
  if (AS == AMDGPUAS::PRIVATE_ADDRESS)
 | 
						|
    return SIAtomicAddrSpace::SCRATCH;
 | 
						|
  if (AS == AMDGPUAS::REGION_ADDRESS)
 | 
						|
    return SIAtomicAddrSpace::GDS;
 | 
						|
 | 
						|
  return SIAtomicAddrSpace::OTHER;
 | 
						|
}
 | 
						|
 | 
						|
SIMemOpAccess::SIMemOpAccess(MachineFunction &MF) {
 | 
						|
  MMI = &MF.getMMI().getObjFileInfo<AMDGPUMachineModuleInfo>();
 | 
						|
}
 | 
						|
 | 
						|
Optional<SIMemOpInfo> SIMemOpAccess::constructFromMIWithMMO(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->getNumMemOperands() > 0);
 | 
						|
 | 
						|
  SyncScope::ID SSID = SyncScope::SingleThread;
 | 
						|
  AtomicOrdering Ordering = AtomicOrdering::NotAtomic;
 | 
						|
  AtomicOrdering FailureOrdering = AtomicOrdering::NotAtomic;
 | 
						|
  SIAtomicAddrSpace InstrAddrSpace = SIAtomicAddrSpace::NONE;
 | 
						|
  bool IsNonTemporal = true;
 | 
						|
 | 
						|
  // Validator should check whether or not MMOs cover the entire set of
 | 
						|
  // locations accessed by the memory instruction.
 | 
						|
  for (const auto &MMO : MI->memoperands()) {
 | 
						|
    IsNonTemporal &= MMO->isNonTemporal();
 | 
						|
    InstrAddrSpace |=
 | 
						|
      toSIAtomicAddrSpace(MMO->getPointerInfo().getAddrSpace());
 | 
						|
    AtomicOrdering OpOrdering = MMO->getOrdering();
 | 
						|
    if (OpOrdering != AtomicOrdering::NotAtomic) {
 | 
						|
      const auto &IsSyncScopeInclusion =
 | 
						|
          MMI->isSyncScopeInclusion(SSID, MMO->getSyncScopeID());
 | 
						|
      if (!IsSyncScopeInclusion) {
 | 
						|
        reportUnsupported(MI,
 | 
						|
          "Unsupported non-inclusive atomic synchronization scope");
 | 
						|
        return None;
 | 
						|
      }
 | 
						|
 | 
						|
      SSID = IsSyncScopeInclusion.getValue() ? SSID : MMO->getSyncScopeID();
 | 
						|
      Ordering =
 | 
						|
          isStrongerThan(Ordering, OpOrdering) ?
 | 
						|
              Ordering : MMO->getOrdering();
 | 
						|
      assert(MMO->getFailureOrdering() != AtomicOrdering::Release &&
 | 
						|
             MMO->getFailureOrdering() != AtomicOrdering::AcquireRelease);
 | 
						|
      FailureOrdering =
 | 
						|
          isStrongerThan(FailureOrdering, MMO->getFailureOrdering()) ?
 | 
						|
              FailureOrdering : MMO->getFailureOrdering();
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  SIAtomicScope Scope = SIAtomicScope::NONE;
 | 
						|
  SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
 | 
						|
  bool IsCrossAddressSpaceOrdering = false;
 | 
						|
  if (Ordering != AtomicOrdering::NotAtomic) {
 | 
						|
    auto ScopeOrNone = toSIAtomicScope(SSID, InstrAddrSpace);
 | 
						|
    if (!ScopeOrNone) {
 | 
						|
      reportUnsupported(MI, "Unsupported atomic synchronization scope");
 | 
						|
      return None;
 | 
						|
    }
 | 
						|
    std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
 | 
						|
      ScopeOrNone.getValue();
 | 
						|
    if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) ||
 | 
						|
        ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) {
 | 
						|
      reportUnsupported(MI, "Unsupported atomic address space");
 | 
						|
      return None;
 | 
						|
    }
 | 
						|
  }
 | 
						|
  return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, InstrAddrSpace,
 | 
						|
                     IsCrossAddressSpaceOrdering, FailureOrdering, IsNonTemporal);
 | 
						|
}
 | 
						|
 | 
						|
Optional<SIMemOpInfo> SIMemOpAccess::getLoadInfo(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic);
 | 
						|
 | 
						|
  if (!(MI->mayLoad() && !MI->mayStore()))
 | 
						|
    return None;
 | 
						|
 | 
						|
  // Be conservative if there are no memory operands.
 | 
						|
  if (MI->getNumMemOperands() == 0)
 | 
						|
    return SIMemOpInfo();
 | 
						|
 | 
						|
  return constructFromMIWithMMO(MI);
 | 
						|
}
 | 
						|
 | 
						|
Optional<SIMemOpInfo> SIMemOpAccess::getStoreInfo(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic);
 | 
						|
 | 
						|
  if (!(!MI->mayLoad() && MI->mayStore()))
 | 
						|
    return None;
 | 
						|
 | 
						|
  // Be conservative if there are no memory operands.
 | 
						|
  if (MI->getNumMemOperands() == 0)
 | 
						|
    return SIMemOpInfo();
 | 
						|
 | 
						|
  return constructFromMIWithMMO(MI);
 | 
						|
}
 | 
						|
 | 
						|
Optional<SIMemOpInfo> SIMemOpAccess::getAtomicFenceInfo(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic);
 | 
						|
 | 
						|
  if (MI->getOpcode() != AMDGPU::ATOMIC_FENCE)
 | 
						|
    return None;
 | 
						|
 | 
						|
  AtomicOrdering Ordering =
 | 
						|
    static_cast<AtomicOrdering>(MI->getOperand(0).getImm());
 | 
						|
 | 
						|
  SyncScope::ID SSID = static_cast<SyncScope::ID>(MI->getOperand(1).getImm());
 | 
						|
  auto ScopeOrNone = toSIAtomicScope(SSID, SIAtomicAddrSpace::ATOMIC);
 | 
						|
  if (!ScopeOrNone) {
 | 
						|
    reportUnsupported(MI, "Unsupported atomic synchronization scope");
 | 
						|
    return None;
 | 
						|
  }
 | 
						|
 | 
						|
  SIAtomicScope Scope = SIAtomicScope::NONE;
 | 
						|
  SIAtomicAddrSpace OrderingAddrSpace = SIAtomicAddrSpace::NONE;
 | 
						|
  bool IsCrossAddressSpaceOrdering = false;
 | 
						|
  std::tie(Scope, OrderingAddrSpace, IsCrossAddressSpaceOrdering) =
 | 
						|
    ScopeOrNone.getValue();
 | 
						|
 | 
						|
  if ((OrderingAddrSpace == SIAtomicAddrSpace::NONE) ||
 | 
						|
      ((OrderingAddrSpace & SIAtomicAddrSpace::ATOMIC) != OrderingAddrSpace)) {
 | 
						|
    reportUnsupported(MI, "Unsupported atomic address space");
 | 
						|
    return None;
 | 
						|
  }
 | 
						|
 | 
						|
  return SIMemOpInfo(Ordering, Scope, OrderingAddrSpace, SIAtomicAddrSpace::ATOMIC,
 | 
						|
                     IsCrossAddressSpaceOrdering);
 | 
						|
}
 | 
						|
 | 
						|
Optional<SIMemOpInfo> SIMemOpAccess::getAtomicCmpxchgOrRmwInfo(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic);
 | 
						|
 | 
						|
  if (!(MI->mayLoad() && MI->mayStore()))
 | 
						|
    return None;
 | 
						|
 | 
						|
  // Be conservative if there are no memory operands.
 | 
						|
  if (MI->getNumMemOperands() == 0)
 | 
						|
    return SIMemOpInfo();
 | 
						|
 | 
						|
  return constructFromMIWithMMO(MI);
 | 
						|
}
 | 
						|
 | 
						|
SICacheControl::SICacheControl(const GCNSubtarget &ST) {
 | 
						|
  TII = ST.getInstrInfo();
 | 
						|
  IV = getIsaVersion(ST.getCPU());
 | 
						|
}
 | 
						|
 | 
						|
/* static */
 | 
						|
std::unique_ptr<SICacheControl> SICacheControl::create(const GCNSubtarget &ST) {
 | 
						|
  GCNSubtarget::Generation Generation = ST.getGeneration();
 | 
						|
  if (Generation <= AMDGPUSubtarget::SOUTHERN_ISLANDS)
 | 
						|
    return make_unique<SIGfx6CacheControl>(ST);
 | 
						|
  return make_unique<SIGfx7CacheControl>(ST);
 | 
						|
}
 | 
						|
 | 
						|
bool SIGfx6CacheControl::enableLoadCacheBypass(
 | 
						|
    const MachineBasicBlock::iterator &MI,
 | 
						|
    SIAtomicScope Scope,
 | 
						|
    SIAtomicAddrSpace AddrSpace) const {
 | 
						|
  assert(MI->mayLoad() && !MI->mayStore());
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) {
 | 
						|
    /// TODO: Do not set glc for rmw atomic operations as they
 | 
						|
    /// implicitly bypass the L1 cache.
 | 
						|
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
      Changed |= enableGLCBit(MI);
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // No cache to bypass.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /// The scratch address space does not need the global memory caches
 | 
						|
  /// to be bypassed as all memory operations by the same thread are
 | 
						|
  /// sequentially consistent, and no other thread can access scratch
 | 
						|
  /// memory.
 | 
						|
 | 
						|
  /// Other address spaces do not hava a cache.
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIGfx6CacheControl::enableNonTemporal(
 | 
						|
    const MachineBasicBlock::iterator &MI) const {
 | 
						|
  assert(MI->mayLoad() ^ MI->mayStore());
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  /// TODO: Do not enableGLCBit if rmw atomic.
 | 
						|
  Changed |= enableGLCBit(MI);
 | 
						|
  Changed |= enableSLCBit(MI);
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIGfx6CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI,
 | 
						|
                                               SIAtomicScope Scope,
 | 
						|
                                               SIAtomicAddrSpace AddrSpace,
 | 
						|
                                               Position Pos) const {
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  MachineBasicBlock &MBB = *MI->getParent();
 | 
						|
  DebugLoc DL = MI->getDebugLoc();
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    ++MI;
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) {
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
      BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1));
 | 
						|
      Changed = true;
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // No cache to invalidate.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /// The scratch address space does not need the global memory cache
 | 
						|
  /// to be flushed as all memory operations by the same thread are
 | 
						|
  /// sequentially consistent, and no other thread can access scratch
 | 
						|
  /// memory.
 | 
						|
 | 
						|
  /// Other address spaces do not hava a cache.
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    --MI;
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIGfx6CacheControl::insertWait(MachineBasicBlock::iterator &MI,
 | 
						|
                                    SIAtomicScope Scope,
 | 
						|
                                    SIAtomicAddrSpace AddrSpace,
 | 
						|
                                    SIMemOp Op,
 | 
						|
                                    bool IsCrossAddrSpaceOrdering,
 | 
						|
                                    Position Pos) const {
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  MachineBasicBlock &MBB = *MI->getParent();
 | 
						|
  DebugLoc DL = MI->getDebugLoc();
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    ++MI;
 | 
						|
 | 
						|
  bool VMCnt = false;
 | 
						|
  bool LGKMCnt = false;
 | 
						|
  bool EXPCnt = false;
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) {
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
      VMCnt = true;
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // The L1 cache keeps all memory operations in order for
 | 
						|
      // wavefronts in the same work-group.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::LDS) != SIAtomicAddrSpace::NONE) {
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
      // If no cross address space ordering then an LDS waitcnt is not
 | 
						|
      // needed as LDS operations for all waves are executed in a
 | 
						|
      // total global ordering as observed by all waves. Required if
 | 
						|
      // also synchronizing with global/GDS memory as LDS operations
 | 
						|
      // could be reordered with respect to later global/GDS memory
 | 
						|
      // operations of the same wave.
 | 
						|
      LGKMCnt = IsCrossAddrSpaceOrdering;
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // The LDS keeps all memory operations in order for
 | 
						|
      // the same wavesfront.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::GDS) != SIAtomicAddrSpace::NONE) {
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
      // If no cross address space ordering then an GDS waitcnt is not
 | 
						|
      // needed as GDS operations for all waves are executed in a
 | 
						|
      // total global ordering as observed by all waves. Required if
 | 
						|
      // also synchronizing with global/LDS memory as GDS operations
 | 
						|
      // could be reordered with respect to later global/LDS memory
 | 
						|
      // operations of the same wave.
 | 
						|
      EXPCnt = IsCrossAddrSpaceOrdering;
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // The GDS keeps all memory operations in order for
 | 
						|
      // the same work-group.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  if (VMCnt || LGKMCnt || EXPCnt) {
 | 
						|
    unsigned WaitCntImmediate =
 | 
						|
      AMDGPU::encodeWaitcnt(IV,
 | 
						|
                            VMCnt ? 0 : getVmcntBitMask(IV),
 | 
						|
                            EXPCnt ? 0 : getExpcntBitMask(IV),
 | 
						|
                            LGKMCnt ? 0 : getLgkmcntBitMask(IV));
 | 
						|
    BuildMI(MBB, MI, DL, TII->get(AMDGPU::S_WAITCNT)).addImm(WaitCntImmediate);
 | 
						|
    Changed = true;
 | 
						|
  }
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    --MI;
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIGfx7CacheControl::insertCacheInvalidate(MachineBasicBlock::iterator &MI,
 | 
						|
                                               SIAtomicScope Scope,
 | 
						|
                                               SIAtomicAddrSpace AddrSpace,
 | 
						|
                                               Position Pos) const {
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  MachineBasicBlock &MBB = *MI->getParent();
 | 
						|
  DebugLoc DL = MI->getDebugLoc();
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    ++MI;
 | 
						|
 | 
						|
  if ((AddrSpace & SIAtomicAddrSpace::GLOBAL) != SIAtomicAddrSpace::NONE) {
 | 
						|
    switch (Scope) {
 | 
						|
    case SIAtomicScope::SYSTEM:
 | 
						|
    case SIAtomicScope::AGENT:
 | 
						|
      BuildMI(MBB, MI, DL, TII->get(AMDGPU::BUFFER_WBINVL1_VOL));
 | 
						|
      Changed = true;
 | 
						|
      break;
 | 
						|
    case SIAtomicScope::WORKGROUP:
 | 
						|
    case SIAtomicScope::WAVEFRONT:
 | 
						|
    case SIAtomicScope::SINGLETHREAD:
 | 
						|
      // No cache to invalidate.
 | 
						|
      break;
 | 
						|
    default:
 | 
						|
      llvm_unreachable("Unsupported synchronization scope");
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  /// The scratch address space does not need the global memory cache
 | 
						|
  /// to be flushed as all memory operations by the same thread are
 | 
						|
  /// sequentially consistent, and no other thread can access scratch
 | 
						|
  /// memory.
 | 
						|
 | 
						|
  /// Other address spaces do not hava a cache.
 | 
						|
 | 
						|
  if (Pos == Position::AFTER)
 | 
						|
    --MI;
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::removeAtomicPseudoMIs() {
 | 
						|
  if (AtomicPseudoMIs.empty())
 | 
						|
    return false;
 | 
						|
 | 
						|
  for (auto &MI : AtomicPseudoMIs)
 | 
						|
    MI->eraseFromParent();
 | 
						|
 | 
						|
  AtomicPseudoMIs.clear();
 | 
						|
  return true;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::expandLoad(const SIMemOpInfo &MOI,
 | 
						|
                                   MachineBasicBlock::iterator &MI) {
 | 
						|
  assert(MI->mayLoad() && !MI->mayStore());
 | 
						|
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  if (MOI.isAtomic()) {
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Monotonic ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) {
 | 
						|
      Changed |= CC->enableLoadCacheBypass(MI, MOI.getScope(),
 | 
						|
                                           MOI.getOrderingAddrSpace());
 | 
						|
    }
 | 
						|
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getOrderingAddrSpace(),
 | 
						|
                                SIMemOp::LOAD | SIMemOp::STORE,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::BEFORE);
 | 
						|
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent) {
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getInstrAddrSpace(),
 | 
						|
                                SIMemOp::LOAD,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::AFTER);
 | 
						|
      Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(),
 | 
						|
                                           MOI.getOrderingAddrSpace(),
 | 
						|
                                           Position::AFTER);
 | 
						|
    }
 | 
						|
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  // Atomic instructions do not have the nontemporal attribute.
 | 
						|
  if (MOI.isNonTemporal()) {
 | 
						|
    Changed |= CC->enableNonTemporal(MI);
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::expandStore(const SIMemOpInfo &MOI,
 | 
						|
                                    MachineBasicBlock::iterator &MI) {
 | 
						|
  assert(!MI->mayLoad() && MI->mayStore());
 | 
						|
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  if (MOI.isAtomic()) {
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Release ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getOrderingAddrSpace(),
 | 
						|
                                SIMemOp::LOAD | SIMemOp::STORE,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::BEFORE);
 | 
						|
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  // Atomic instructions do not have the nontemporal attribute.
 | 
						|
  if (MOI.isNonTemporal()) {
 | 
						|
    Changed |= CC->enableNonTemporal(MI);
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::expandAtomicFence(const SIMemOpInfo &MOI,
 | 
						|
                                          MachineBasicBlock::iterator &MI) {
 | 
						|
  assert(MI->getOpcode() == AMDGPU::ATOMIC_FENCE);
 | 
						|
 | 
						|
  AtomicPseudoMIs.push_back(MI);
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  if (MOI.isAtomic()) {
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::Release ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::AcquireRelease ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
 | 
						|
      /// TODO: This relies on a barrier always generating a waitcnt
 | 
						|
      /// for LDS to ensure it is not reordered with the completion of
 | 
						|
      /// the proceeding LDS operations. If barrier had a memory
 | 
						|
      /// ordering and memory scope, then library does not need to
 | 
						|
      /// generate a fence. Could add support in this file for
 | 
						|
      /// barrier. SIInsertWaitcnt.cpp could then stop unconditionally
 | 
						|
      /// adding waitcnt before a S_BARRIER.
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getOrderingAddrSpace(),
 | 
						|
                                SIMemOp::LOAD | SIMemOp::STORE,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::BEFORE);
 | 
						|
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::AcquireRelease ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent)
 | 
						|
      Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(),
 | 
						|
                                           MOI.getOrderingAddrSpace(),
 | 
						|
                                           Position::BEFORE);
 | 
						|
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::expandAtomicCmpxchgOrRmw(const SIMemOpInfo &MOI,
 | 
						|
  MachineBasicBlock::iterator &MI) {
 | 
						|
  assert(MI->mayLoad() && MI->mayStore());
 | 
						|
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  if (MOI.isAtomic()) {
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Release ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::AcquireRelease ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent ||
 | 
						|
        MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent)
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getOrderingAddrSpace(),
 | 
						|
                                SIMemOp::LOAD | SIMemOp::STORE,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::BEFORE);
 | 
						|
 | 
						|
    if (MOI.getOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::AcquireRelease ||
 | 
						|
        MOI.getOrdering() == AtomicOrdering::SequentiallyConsistent ||
 | 
						|
        MOI.getFailureOrdering() == AtomicOrdering::Acquire ||
 | 
						|
        MOI.getFailureOrdering() == AtomicOrdering::SequentiallyConsistent) {
 | 
						|
      Changed |= CC->insertWait(MI, MOI.getScope(),
 | 
						|
                                MOI.getOrderingAddrSpace(),
 | 
						|
                                isAtomicRet(*MI) ? SIMemOp::LOAD :
 | 
						|
                                                   SIMemOp::STORE,
 | 
						|
                                MOI.getIsCrossAddressSpaceOrdering(),
 | 
						|
                                Position::AFTER);
 | 
						|
      Changed |= CC->insertCacheInvalidate(MI, MOI.getScope(),
 | 
						|
                                           MOI.getOrderingAddrSpace(),
 | 
						|
                                           Position::AFTER);
 | 
						|
    }
 | 
						|
 | 
						|
    return Changed;
 | 
						|
  }
 | 
						|
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
bool SIMemoryLegalizer::runOnMachineFunction(MachineFunction &MF) {
 | 
						|
  bool Changed = false;
 | 
						|
 | 
						|
  SIMemOpAccess MOA(MF);
 | 
						|
  CC = SICacheControl::create(MF.getSubtarget<GCNSubtarget>());
 | 
						|
 | 
						|
  for (auto &MBB : MF) {
 | 
						|
    for (auto MI = MBB.begin(); MI != MBB.end(); ++MI) {
 | 
						|
      if (!(MI->getDesc().TSFlags & SIInstrFlags::maybeAtomic))
 | 
						|
        continue;
 | 
						|
 | 
						|
      if (const auto &MOI = MOA.getLoadInfo(MI))
 | 
						|
        Changed |= expandLoad(MOI.getValue(), MI);
 | 
						|
      else if (const auto &MOI = MOA.getStoreInfo(MI))
 | 
						|
        Changed |= expandStore(MOI.getValue(), MI);
 | 
						|
      else if (const auto &MOI = MOA.getAtomicFenceInfo(MI))
 | 
						|
        Changed |= expandAtomicFence(MOI.getValue(), MI);
 | 
						|
      else if (const auto &MOI = MOA.getAtomicCmpxchgOrRmwInfo(MI))
 | 
						|
        Changed |= expandAtomicCmpxchgOrRmw(MOI.getValue(), MI);
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  Changed |= removeAtomicPseudoMIs();
 | 
						|
  return Changed;
 | 
						|
}
 | 
						|
 | 
						|
INITIALIZE_PASS(SIMemoryLegalizer, DEBUG_TYPE, PASS_NAME, false, false)
 | 
						|
 | 
						|
char SIMemoryLegalizer::ID = 0;
 | 
						|
char &llvm::SIMemoryLegalizerID = SIMemoryLegalizer::ID;
 | 
						|
 | 
						|
FunctionPass *llvm::createSIMemoryLegalizerPass() {
 | 
						|
  return new SIMemoryLegalizer();
 | 
						|
}
 |