| //===-- NVPTXISelDAGToDAG.cpp - A dag to dag inst selector for NVPTX ------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file defines an instruction selector for the NVPTX target. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "NVPTXISelDAGToDAG.h" |
| #include "NVPTXUtilities.h" |
| #include "llvm/Analysis/ValueTracking.h" |
| #include "llvm/IR/GlobalValue.h" |
| #include "llvm/IR/Instructions.h" |
| #include "llvm/Support/AtomicOrdering.h" |
| #include "llvm/Support/CommandLine.h" |
| #include "llvm/Support/Debug.h" |
| #include "llvm/Support/ErrorHandling.h" |
| #include "llvm/Support/raw_ostream.h" |
| #include "llvm/Target/TargetIntrinsicInfo.h" |
| |
| using namespace llvm; |
| |
| #define DEBUG_TYPE "nvptx-isel" |
| |
| /// createNVPTXISelDag - This pass converts a legalized DAG into a |
| /// NVPTX-specific DAG, ready for instruction scheduling. |
| FunctionPass *llvm::createNVPTXISelDag(NVPTXTargetMachine &TM, |
| llvm::CodeGenOpt::Level OptLevel) { |
| return new NVPTXDAGToDAGISel(TM, OptLevel); |
| } |
| |
| NVPTXDAGToDAGISel::NVPTXDAGToDAGISel(NVPTXTargetMachine &tm, |
| CodeGenOpt::Level OptLevel) |
| : SelectionDAGISel(tm, OptLevel), TM(tm) { |
| doMulWide = (OptLevel > 0); |
| } |
| |
| bool NVPTXDAGToDAGISel::runOnMachineFunction(MachineFunction &MF) { |
| Subtarget = &static_cast<const NVPTXSubtarget &>(MF.getSubtarget()); |
| return SelectionDAGISel::runOnMachineFunction(MF); |
| } |
| |
| int NVPTXDAGToDAGISel::getDivF32Level() const { |
| return Subtarget->getTargetLowering()->getDivF32Level(); |
| } |
| |
| bool NVPTXDAGToDAGISel::usePrecSqrtF32() const { |
| return Subtarget->getTargetLowering()->usePrecSqrtF32(); |
| } |
| |
| bool NVPTXDAGToDAGISel::useF32FTZ() const { |
| return Subtarget->getTargetLowering()->useF32FTZ(*MF); |
| } |
| |
| bool NVPTXDAGToDAGISel::allowFMA() const { |
| const NVPTXTargetLowering *TL = Subtarget->getTargetLowering(); |
| return TL->allowFMA(*MF, OptLevel); |
| } |
| |
| bool NVPTXDAGToDAGISel::allowUnsafeFPMath() const { |
| const NVPTXTargetLowering *TL = Subtarget->getTargetLowering(); |
| return TL->allowUnsafeFPMath(*MF); |
| } |
| |
| bool NVPTXDAGToDAGISel::useShortPointers() const { |
| return TM.useShortPointers(); |
| } |
| |
| /// Select - Select instructions not customized! Used for |
| /// expanded, promoted and normal instructions. |
| void NVPTXDAGToDAGISel::Select(SDNode *N) { |
| |
| if (N->isMachineOpcode()) { |
| N->setNodeId(-1); |
| return; // Already selected. |
| } |
| |
| switch (N->getOpcode()) { |
| case ISD::LOAD: |
| case ISD::ATOMIC_LOAD: |
| if (tryLoad(N)) |
| return; |
| break; |
| case ISD::STORE: |
| case ISD::ATOMIC_STORE: |
| if (tryStore(N)) |
| return; |
| break; |
| case ISD::EXTRACT_VECTOR_ELT: |
| if (tryEXTRACT_VECTOR_ELEMENT(N)) |
| return; |
| break; |
| case NVPTXISD::SETP_F16X2: |
| SelectSETP_F16X2(N); |
| return; |
| |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LoadV4: |
| if (tryLoadVector(N)) |
| return; |
| break; |
| case NVPTXISD::LDGV2: |
| case NVPTXISD::LDGV4: |
| case NVPTXISD::LDUV2: |
| case NVPTXISD::LDUV4: |
| if (tryLDGLDU(N)) |
| return; |
| break; |
| case NVPTXISD::StoreV2: |
| case NVPTXISD::StoreV4: |
| if (tryStoreVector(N)) |
| return; |
| break; |
| case NVPTXISD::LoadParam: |
| case NVPTXISD::LoadParamV2: |
| case NVPTXISD::LoadParamV4: |
| if (tryLoadParam(N)) |
| return; |
| break; |
| case NVPTXISD::StoreRetval: |
| case NVPTXISD::StoreRetvalV2: |
| case NVPTXISD::StoreRetvalV4: |
| if (tryStoreRetval(N)) |
| return; |
| break; |
| case NVPTXISD::StoreParam: |
| case NVPTXISD::StoreParamV2: |
| case NVPTXISD::StoreParamV4: |
| case NVPTXISD::StoreParamS32: |
| case NVPTXISD::StoreParamU32: |
| if (tryStoreParam(N)) |
| return; |
| break; |
| case ISD::INTRINSIC_WO_CHAIN: |
| if (tryIntrinsicNoChain(N)) |
| return; |
| break; |
| case ISD::INTRINSIC_W_CHAIN: |
| if (tryIntrinsicChain(N)) |
| return; |
| break; |
| case NVPTXISD::Tex1DFloatS32: |
| case NVPTXISD::Tex1DFloatFloat: |
| case NVPTXISD::Tex1DFloatFloatLevel: |
| case NVPTXISD::Tex1DFloatFloatGrad: |
| case NVPTXISD::Tex1DS32S32: |
| case NVPTXISD::Tex1DS32Float: |
| case NVPTXISD::Tex1DS32FloatLevel: |
| case NVPTXISD::Tex1DS32FloatGrad: |
| case NVPTXISD::Tex1DU32S32: |
| case NVPTXISD::Tex1DU32Float: |
| case NVPTXISD::Tex1DU32FloatLevel: |
| case NVPTXISD::Tex1DU32FloatGrad: |
| case NVPTXISD::Tex1DArrayFloatS32: |
| case NVPTXISD::Tex1DArrayFloatFloat: |
| case NVPTXISD::Tex1DArrayFloatFloatLevel: |
| case NVPTXISD::Tex1DArrayFloatFloatGrad: |
| case NVPTXISD::Tex1DArrayS32S32: |
| case NVPTXISD::Tex1DArrayS32Float: |
| case NVPTXISD::Tex1DArrayS32FloatLevel: |
| case NVPTXISD::Tex1DArrayS32FloatGrad: |
| case NVPTXISD::Tex1DArrayU32S32: |
| case NVPTXISD::Tex1DArrayU32Float: |
| case NVPTXISD::Tex1DArrayU32FloatLevel: |
| case NVPTXISD::Tex1DArrayU32FloatGrad: |
| case NVPTXISD::Tex2DFloatS32: |
| case NVPTXISD::Tex2DFloatFloat: |
| case NVPTXISD::Tex2DFloatFloatLevel: |
| case NVPTXISD::Tex2DFloatFloatGrad: |
| case NVPTXISD::Tex2DS32S32: |
| case NVPTXISD::Tex2DS32Float: |
| case NVPTXISD::Tex2DS32FloatLevel: |
| case NVPTXISD::Tex2DS32FloatGrad: |
| case NVPTXISD::Tex2DU32S32: |
| case NVPTXISD::Tex2DU32Float: |
| case NVPTXISD::Tex2DU32FloatLevel: |
| case NVPTXISD::Tex2DU32FloatGrad: |
| case NVPTXISD::Tex2DArrayFloatS32: |
| case NVPTXISD::Tex2DArrayFloatFloat: |
| case NVPTXISD::Tex2DArrayFloatFloatLevel: |
| case NVPTXISD::Tex2DArrayFloatFloatGrad: |
| case NVPTXISD::Tex2DArrayS32S32: |
| case NVPTXISD::Tex2DArrayS32Float: |
| case NVPTXISD::Tex2DArrayS32FloatLevel: |
| case NVPTXISD::Tex2DArrayS32FloatGrad: |
| case NVPTXISD::Tex2DArrayU32S32: |
| case NVPTXISD::Tex2DArrayU32Float: |
| case NVPTXISD::Tex2DArrayU32FloatLevel: |
| case NVPTXISD::Tex2DArrayU32FloatGrad: |
| case NVPTXISD::Tex3DFloatS32: |
| case NVPTXISD::Tex3DFloatFloat: |
| case NVPTXISD::Tex3DFloatFloatLevel: |
| case NVPTXISD::Tex3DFloatFloatGrad: |
| case NVPTXISD::Tex3DS32S32: |
| case NVPTXISD::Tex3DS32Float: |
| case NVPTXISD::Tex3DS32FloatLevel: |
| case NVPTXISD::Tex3DS32FloatGrad: |
| case NVPTXISD::Tex3DU32S32: |
| case NVPTXISD::Tex3DU32Float: |
| case NVPTXISD::Tex3DU32FloatLevel: |
| case NVPTXISD::Tex3DU32FloatGrad: |
| case NVPTXISD::TexCubeFloatFloat: |
| case NVPTXISD::TexCubeFloatFloatLevel: |
| case NVPTXISD::TexCubeS32Float: |
| case NVPTXISD::TexCubeS32FloatLevel: |
| case NVPTXISD::TexCubeU32Float: |
| case NVPTXISD::TexCubeU32FloatLevel: |
| case NVPTXISD::TexCubeArrayFloatFloat: |
| case NVPTXISD::TexCubeArrayFloatFloatLevel: |
| case NVPTXISD::TexCubeArrayS32Float: |
| case NVPTXISD::TexCubeArrayS32FloatLevel: |
| case NVPTXISD::TexCubeArrayU32Float: |
| case NVPTXISD::TexCubeArrayU32FloatLevel: |
| case NVPTXISD::Tld4R2DFloatFloat: |
| case NVPTXISD::Tld4G2DFloatFloat: |
| case NVPTXISD::Tld4B2DFloatFloat: |
| case NVPTXISD::Tld4A2DFloatFloat: |
| case NVPTXISD::Tld4R2DS64Float: |
| case NVPTXISD::Tld4G2DS64Float: |
| case NVPTXISD::Tld4B2DS64Float: |
| case NVPTXISD::Tld4A2DS64Float: |
| case NVPTXISD::Tld4R2DU64Float: |
| case NVPTXISD::Tld4G2DU64Float: |
| case NVPTXISD::Tld4B2DU64Float: |
| case NVPTXISD::Tld4A2DU64Float: |
| case NVPTXISD::TexUnified1DFloatS32: |
| case NVPTXISD::TexUnified1DFloatFloat: |
| case NVPTXISD::TexUnified1DFloatFloatLevel: |
| case NVPTXISD::TexUnified1DFloatFloatGrad: |
| case NVPTXISD::TexUnified1DS32S32: |
| case NVPTXISD::TexUnified1DS32Float: |
| case NVPTXISD::TexUnified1DS32FloatLevel: |
| case NVPTXISD::TexUnified1DS32FloatGrad: |
| case NVPTXISD::TexUnified1DU32S32: |
| case NVPTXISD::TexUnified1DU32Float: |
| case NVPTXISD::TexUnified1DU32FloatLevel: |
| case NVPTXISD::TexUnified1DU32FloatGrad: |
| case NVPTXISD::TexUnified1DArrayFloatS32: |
| case NVPTXISD::TexUnified1DArrayFloatFloat: |
| case NVPTXISD::TexUnified1DArrayFloatFloatLevel: |
| case NVPTXISD::TexUnified1DArrayFloatFloatGrad: |
| case NVPTXISD::TexUnified1DArrayS32S32: |
| case NVPTXISD::TexUnified1DArrayS32Float: |
| case NVPTXISD::TexUnified1DArrayS32FloatLevel: |
| case NVPTXISD::TexUnified1DArrayS32FloatGrad: |
| case NVPTXISD::TexUnified1DArrayU32S32: |
| case NVPTXISD::TexUnified1DArrayU32Float: |
| case NVPTXISD::TexUnified1DArrayU32FloatLevel: |
| case NVPTXISD::TexUnified1DArrayU32FloatGrad: |
| case NVPTXISD::TexUnified2DFloatS32: |
| case NVPTXISD::TexUnified2DFloatFloat: |
| case NVPTXISD::TexUnified2DFloatFloatLevel: |
| case NVPTXISD::TexUnified2DFloatFloatGrad: |
| case NVPTXISD::TexUnified2DS32S32: |
| case NVPTXISD::TexUnified2DS32Float: |
| case NVPTXISD::TexUnified2DS32FloatLevel: |
| case NVPTXISD::TexUnified2DS32FloatGrad: |
| case NVPTXISD::TexUnified2DU32S32: |
| case NVPTXISD::TexUnified2DU32Float: |
| case NVPTXISD::TexUnified2DU32FloatLevel: |
| case NVPTXISD::TexUnified2DU32FloatGrad: |
| case NVPTXISD::TexUnified2DArrayFloatS32: |
| case NVPTXISD::TexUnified2DArrayFloatFloat: |
| case NVPTXISD::TexUnified2DArrayFloatFloatLevel: |
| case NVPTXISD::TexUnified2DArrayFloatFloatGrad: |
| case NVPTXISD::TexUnified2DArrayS32S32: |
| case NVPTXISD::TexUnified2DArrayS32Float: |
| case NVPTXISD::TexUnified2DArrayS32FloatLevel: |
| case NVPTXISD::TexUnified2DArrayS32FloatGrad: |
| case NVPTXISD::TexUnified2DArrayU32S32: |
| case NVPTXISD::TexUnified2DArrayU32Float: |
| case NVPTXISD::TexUnified2DArrayU32FloatLevel: |
| case NVPTXISD::TexUnified2DArrayU32FloatGrad: |
| case NVPTXISD::TexUnified3DFloatS32: |
| case NVPTXISD::TexUnified3DFloatFloat: |
| case NVPTXISD::TexUnified3DFloatFloatLevel: |
| case NVPTXISD::TexUnified3DFloatFloatGrad: |
| case NVPTXISD::TexUnified3DS32S32: |
| case NVPTXISD::TexUnified3DS32Float: |
| case NVPTXISD::TexUnified3DS32FloatLevel: |
| case NVPTXISD::TexUnified3DS32FloatGrad: |
| case NVPTXISD::TexUnified3DU32S32: |
| case NVPTXISD::TexUnified3DU32Float: |
| case NVPTXISD::TexUnified3DU32FloatLevel: |
| case NVPTXISD::TexUnified3DU32FloatGrad: |
| case NVPTXISD::TexUnifiedCubeFloatFloat: |
| case NVPTXISD::TexUnifiedCubeFloatFloatLevel: |
| case NVPTXISD::TexUnifiedCubeS32Float: |
| case NVPTXISD::TexUnifiedCubeS32FloatLevel: |
| case NVPTXISD::TexUnifiedCubeU32Float: |
| case NVPTXISD::TexUnifiedCubeU32FloatLevel: |
| case NVPTXISD::TexUnifiedCubeArrayFloatFloat: |
| case NVPTXISD::TexUnifiedCubeArrayFloatFloatLevel: |
| case NVPTXISD::TexUnifiedCubeArrayS32Float: |
| case NVPTXISD::TexUnifiedCubeArrayS32FloatLevel: |
| case NVPTXISD::TexUnifiedCubeArrayU32Float: |
| case NVPTXISD::TexUnifiedCubeArrayU32FloatLevel: |
| case NVPTXISD::Tld4UnifiedR2DFloatFloat: |
| case NVPTXISD::Tld4UnifiedG2DFloatFloat: |
| case NVPTXISD::Tld4UnifiedB2DFloatFloat: |
| case NVPTXISD::Tld4UnifiedA2DFloatFloat: |
| case NVPTXISD::Tld4UnifiedR2DS64Float: |
| case NVPTXISD::Tld4UnifiedG2DS64Float: |
| case NVPTXISD::Tld4UnifiedB2DS64Float: |
| case NVPTXISD::Tld4UnifiedA2DS64Float: |
| case NVPTXISD::Tld4UnifiedR2DU64Float: |
| case NVPTXISD::Tld4UnifiedG2DU64Float: |
| case NVPTXISD::Tld4UnifiedB2DU64Float: |
| case NVPTXISD::Tld4UnifiedA2DU64Float: |
| if (tryTextureIntrinsic(N)) |
| return; |
| break; |
| case NVPTXISD::Suld1DI8Clamp: |
| case NVPTXISD::Suld1DI16Clamp: |
| case NVPTXISD::Suld1DI32Clamp: |
| case NVPTXISD::Suld1DI64Clamp: |
| case NVPTXISD::Suld1DV2I8Clamp: |
| case NVPTXISD::Suld1DV2I16Clamp: |
| case NVPTXISD::Suld1DV2I32Clamp: |
| case NVPTXISD::Suld1DV2I64Clamp: |
| case NVPTXISD::Suld1DV4I8Clamp: |
| case NVPTXISD::Suld1DV4I16Clamp: |
| case NVPTXISD::Suld1DV4I32Clamp: |
| case NVPTXISD::Suld1DArrayI8Clamp: |
| case NVPTXISD::Suld1DArrayI16Clamp: |
| case NVPTXISD::Suld1DArrayI32Clamp: |
| case NVPTXISD::Suld1DArrayI64Clamp: |
| case NVPTXISD::Suld1DArrayV2I8Clamp: |
| case NVPTXISD::Suld1DArrayV2I16Clamp: |
| case NVPTXISD::Suld1DArrayV2I32Clamp: |
| case NVPTXISD::Suld1DArrayV2I64Clamp: |
| case NVPTXISD::Suld1DArrayV4I8Clamp: |
| case NVPTXISD::Suld1DArrayV4I16Clamp: |
| case NVPTXISD::Suld1DArrayV4I32Clamp: |
| case NVPTXISD::Suld2DI8Clamp: |
| case NVPTXISD::Suld2DI16Clamp: |
| case NVPTXISD::Suld2DI32Clamp: |
| case NVPTXISD::Suld2DI64Clamp: |
| case NVPTXISD::Suld2DV2I8Clamp: |
| case NVPTXISD::Suld2DV2I16Clamp: |
| case NVPTXISD::Suld2DV2I32Clamp: |
| case NVPTXISD::Suld2DV2I64Clamp: |
| case NVPTXISD::Suld2DV4I8Clamp: |
| case NVPTXISD::Suld2DV4I16Clamp: |
| case NVPTXISD::Suld2DV4I32Clamp: |
| case NVPTXISD::Suld2DArrayI8Clamp: |
| case NVPTXISD::Suld2DArrayI16Clamp: |
| case NVPTXISD::Suld2DArrayI32Clamp: |
| case NVPTXISD::Suld2DArrayI64Clamp: |
| case NVPTXISD::Suld2DArrayV2I8Clamp: |
| case NVPTXISD::Suld2DArrayV2I16Clamp: |
| case NVPTXISD::Suld2DArrayV2I32Clamp: |
| case NVPTXISD::Suld2DArrayV2I64Clamp: |
| case NVPTXISD::Suld2DArrayV4I8Clamp: |
| case NVPTXISD::Suld2DArrayV4I16Clamp: |
| case NVPTXISD::Suld2DArrayV4I32Clamp: |
| case NVPTXISD::Suld3DI8Clamp: |
| case NVPTXISD::Suld3DI16Clamp: |
| case NVPTXISD::Suld3DI32Clamp: |
| case NVPTXISD::Suld3DI64Clamp: |
| case NVPTXISD::Suld3DV2I8Clamp: |
| case NVPTXISD::Suld3DV2I16Clamp: |
| case NVPTXISD::Suld3DV2I32Clamp: |
| case NVPTXISD::Suld3DV2I64Clamp: |
| case NVPTXISD::Suld3DV4I8Clamp: |
| case NVPTXISD::Suld3DV4I16Clamp: |
| case NVPTXISD::Suld3DV4I32Clamp: |
| case NVPTXISD::Suld1DI8Trap: |
| case NVPTXISD::Suld1DI16Trap: |
| case NVPTXISD::Suld1DI32Trap: |
| case NVPTXISD::Suld1DI64Trap: |
| case NVPTXISD::Suld1DV2I8Trap: |
| case NVPTXISD::Suld1DV2I16Trap: |
| case NVPTXISD::Suld1DV2I32Trap: |
| case NVPTXISD::Suld1DV2I64Trap: |
| case NVPTXISD::Suld1DV4I8Trap: |
| case NVPTXISD::Suld1DV4I16Trap: |
| case NVPTXISD::Suld1DV4I32Trap: |
| case NVPTXISD::Suld1DArrayI8Trap: |
| case NVPTXISD::Suld1DArrayI16Trap: |
| case NVPTXISD::Suld1DArrayI32Trap: |
| case NVPTXISD::Suld1DArrayI64Trap: |
| case NVPTXISD::Suld1DArrayV2I8Trap: |
| case NVPTXISD::Suld1DArrayV2I16Trap: |
| case NVPTXISD::Suld1DArrayV2I32Trap: |
| case NVPTXISD::Suld1DArrayV2I64Trap: |
| case NVPTXISD::Suld1DArrayV4I8Trap: |
| case NVPTXISD::Suld1DArrayV4I16Trap: |
| case NVPTXISD::Suld1DArrayV4I32Trap: |
| case NVPTXISD::Suld2DI8Trap: |
| case NVPTXISD::Suld2DI16Trap: |
| case NVPTXISD::Suld2DI32Trap: |
| case NVPTXISD::Suld2DI64Trap: |
| case NVPTXISD::Suld2DV2I8Trap: |
| case NVPTXISD::Suld2DV2I16Trap: |
| case NVPTXISD::Suld2DV2I32Trap: |
| case NVPTXISD::Suld2DV2I64Trap: |
| case NVPTXISD::Suld2DV4I8Trap: |
| case NVPTXISD::Suld2DV4I16Trap: |
| case NVPTXISD::Suld2DV4I32Trap: |
| case NVPTXISD::Suld2DArrayI8Trap: |
| case NVPTXISD::Suld2DArrayI16Trap: |
| case NVPTXISD::Suld2DArrayI32Trap: |
| case NVPTXISD::Suld2DArrayI64Trap: |
| case NVPTXISD::Suld2DArrayV2I8Trap: |
| case NVPTXISD::Suld2DArrayV2I16Trap: |
| case NVPTXISD::Suld2DArrayV2I32Trap: |
| case NVPTXISD::Suld2DArrayV2I64Trap: |
| case NVPTXISD::Suld2DArrayV4I8Trap: |
| case NVPTXISD::Suld2DArrayV4I16Trap: |
| case NVPTXISD::Suld2DArrayV4I32Trap: |
| case NVPTXISD::Suld3DI8Trap: |
| case NVPTXISD::Suld3DI16Trap: |
| case NVPTXISD::Suld3DI32Trap: |
| case NVPTXISD::Suld3DI64Trap: |
| case NVPTXISD::Suld3DV2I8Trap: |
| case NVPTXISD::Suld3DV2I16Trap: |
| case NVPTXISD::Suld3DV2I32Trap: |
| case NVPTXISD::Suld3DV2I64Trap: |
| case NVPTXISD::Suld3DV4I8Trap: |
| case NVPTXISD::Suld3DV4I16Trap: |
| case NVPTXISD::Suld3DV4I32Trap: |
| case NVPTXISD::Suld1DI8Zero: |
| case NVPTXISD::Suld1DI16Zero: |
| case NVPTXISD::Suld1DI32Zero: |
| case NVPTXISD::Suld1DI64Zero: |
| case NVPTXISD::Suld1DV2I8Zero: |
| case NVPTXISD::Suld1DV2I16Zero: |
| case NVPTXISD::Suld1DV2I32Zero: |
| case NVPTXISD::Suld1DV2I64Zero: |
| case NVPTXISD::Suld1DV4I8Zero: |
| case NVPTXISD::Suld1DV4I16Zero: |
| case NVPTXISD::Suld1DV4I32Zero: |
| case NVPTXISD::Suld1DArrayI8Zero: |
| case NVPTXISD::Suld1DArrayI16Zero: |
| case NVPTXISD::Suld1DArrayI32Zero: |
| case NVPTXISD::Suld1DArrayI64Zero: |
| case NVPTXISD::Suld1DArrayV2I8Zero: |
| case NVPTXISD::Suld1DArrayV2I16Zero: |
| case NVPTXISD::Suld1DArrayV2I32Zero: |
| case NVPTXISD::Suld1DArrayV2I64Zero: |
| case NVPTXISD::Suld1DArrayV4I8Zero: |
| case NVPTXISD::Suld1DArrayV4I16Zero: |
| case NVPTXISD::Suld1DArrayV4I32Zero: |
| case NVPTXISD::Suld2DI8Zero: |
| case NVPTXISD::Suld2DI16Zero: |
| case NVPTXISD::Suld2DI32Zero: |
| case NVPTXISD::Suld2DI64Zero: |
| case NVPTXISD::Suld2DV2I8Zero: |
| case NVPTXISD::Suld2DV2I16Zero: |
| case NVPTXISD::Suld2DV2I32Zero: |
| case NVPTXISD::Suld2DV2I64Zero: |
| case NVPTXISD::Suld2DV4I8Zero: |
| case NVPTXISD::Suld2DV4I16Zero: |
| case NVPTXISD::Suld2DV4I32Zero: |
| case NVPTXISD::Suld2DArrayI8Zero: |
| case NVPTXISD::Suld2DArrayI16Zero: |
| case NVPTXISD::Suld2DArrayI32Zero: |
| case NVPTXISD::Suld2DArrayI64Zero: |
| case NVPTXISD::Suld2DArrayV2I8Zero: |
| case NVPTXISD::Suld2DArrayV2I16Zero: |
| case NVPTXISD::Suld2DArrayV2I32Zero: |
| case NVPTXISD::Suld2DArrayV2I64Zero: |
| case NVPTXISD::Suld2DArrayV4I8Zero: |
| case NVPTXISD::Suld2DArrayV4I16Zero: |
| case NVPTXISD::Suld2DArrayV4I32Zero: |
| case NVPTXISD::Suld3DI8Zero: |
| case NVPTXISD::Suld3DI16Zero: |
| case NVPTXISD::Suld3DI32Zero: |
| case NVPTXISD::Suld3DI64Zero: |
| case NVPTXISD::Suld3DV2I8Zero: |
| case NVPTXISD::Suld3DV2I16Zero: |
| case NVPTXISD::Suld3DV2I32Zero: |
| case NVPTXISD::Suld3DV2I64Zero: |
| case NVPTXISD::Suld3DV4I8Zero: |
| case NVPTXISD::Suld3DV4I16Zero: |
| case NVPTXISD::Suld3DV4I32Zero: |
| if (trySurfaceIntrinsic(N)) |
| return; |
| break; |
| case ISD::AND: |
| case ISD::SRA: |
| case ISD::SRL: |
| // Try to select BFE |
| if (tryBFE(N)) |
| return; |
| break; |
| case ISD::ADDRSPACECAST: |
| SelectAddrSpaceCast(N); |
| return; |
| case ISD::ConstantFP: |
| if (tryConstantFP16(N)) |
| return; |
| break; |
| default: |
| break; |
| } |
| SelectCode(N); |
| } |
| |
| bool NVPTXDAGToDAGISel::tryIntrinsicChain(SDNode *N) { |
| unsigned IID = cast<ConstantSDNode>(N->getOperand(1))->getZExtValue(); |
| switch (IID) { |
| default: |
| return false; |
| case Intrinsic::nvvm_ldg_global_f: |
| case Intrinsic::nvvm_ldg_global_i: |
| case Intrinsic::nvvm_ldg_global_p: |
| case Intrinsic::nvvm_ldu_global_f: |
| case Intrinsic::nvvm_ldu_global_i: |
| case Intrinsic::nvvm_ldu_global_p: |
| return tryLDGLDU(N); |
| } |
| } |
| |
| // There's no way to specify FP16 immediates in .f16 ops, so we have to |
| // load them into an .f16 register first. |
| bool NVPTXDAGToDAGISel::tryConstantFP16(SDNode *N) { |
| if (N->getValueType(0) != MVT::f16) |
| return false; |
| SDValue Val = CurDAG->getTargetConstantFP( |
| cast<ConstantFPSDNode>(N)->getValueAPF(), SDLoc(N), MVT::f16); |
| SDNode *LoadConstF16 = |
| CurDAG->getMachineNode(NVPTX::LOAD_CONST_F16, SDLoc(N), MVT::f16, Val); |
| ReplaceNode(N, LoadConstF16); |
| return true; |
| } |
| |
| // Map ISD:CONDCODE value to appropriate CmpMode expected by |
| // NVPTXInstPrinter::printCmpMode() |
| static unsigned getPTXCmpMode(const CondCodeSDNode &CondCode, bool FTZ) { |
| using NVPTX::PTXCmpMode::CmpMode; |
| unsigned PTXCmpMode = [](ISD::CondCode CC) { |
| switch (CC) { |
| default: |
| llvm_unreachable("Unexpected condition code."); |
| case ISD::SETOEQ: |
| return CmpMode::EQ; |
| case ISD::SETOGT: |
| return CmpMode::GT; |
| case ISD::SETOGE: |
| return CmpMode::GE; |
| case ISD::SETOLT: |
| return CmpMode::LT; |
| case ISD::SETOLE: |
| return CmpMode::LE; |
| case ISD::SETONE: |
| return CmpMode::NE; |
| case ISD::SETO: |
| return CmpMode::NUM; |
| case ISD::SETUO: |
| return CmpMode::NotANumber; |
| case ISD::SETUEQ: |
| return CmpMode::EQU; |
| case ISD::SETUGT: |
| return CmpMode::GTU; |
| case ISD::SETUGE: |
| return CmpMode::GEU; |
| case ISD::SETULT: |
| return CmpMode::LTU; |
| case ISD::SETULE: |
| return CmpMode::LEU; |
| case ISD::SETUNE: |
| return CmpMode::NEU; |
| case ISD::SETEQ: |
| return CmpMode::EQ; |
| case ISD::SETGT: |
| return CmpMode::GT; |
| case ISD::SETGE: |
| return CmpMode::GE; |
| case ISD::SETLT: |
| return CmpMode::LT; |
| case ISD::SETLE: |
| return CmpMode::LE; |
| case ISD::SETNE: |
| return CmpMode::NE; |
| } |
| }(CondCode.get()); |
| |
| if (FTZ) |
| PTXCmpMode |= NVPTX::PTXCmpMode::FTZ_FLAG; |
| |
| return PTXCmpMode; |
| } |
| |
| bool NVPTXDAGToDAGISel::SelectSETP_F16X2(SDNode *N) { |
| unsigned PTXCmpMode = |
| getPTXCmpMode(*cast<CondCodeSDNode>(N->getOperand(2)), useF32FTZ()); |
| SDLoc DL(N); |
| SDNode *SetP = CurDAG->getMachineNode( |
| NVPTX::SETP_f16x2rr, DL, MVT::i1, MVT::i1, N->getOperand(0), |
| N->getOperand(1), CurDAG->getTargetConstant(PTXCmpMode, DL, MVT::i32)); |
| ReplaceNode(N, SetP); |
| return true; |
| } |
| |
| // Find all instances of extract_vector_elt that use this v2f16 vector |
| // and coalesce them into a scattering move instruction. |
| bool NVPTXDAGToDAGISel::tryEXTRACT_VECTOR_ELEMENT(SDNode *N) { |
| SDValue Vector = N->getOperand(0); |
| |
| // We only care about f16x2 as it's the only real vector type we |
| // need to deal with. |
| if (Vector.getSimpleValueType() != MVT::v2f16) |
| return false; |
| |
| // Find and record all uses of this vector that extract element 0 or 1. |
| SmallVector<SDNode *, 4> E0, E1; |
| for (const auto &U : Vector.getNode()->uses()) { |
| if (U->getOpcode() != ISD::EXTRACT_VECTOR_ELT) |
| continue; |
| if (U->getOperand(0) != Vector) |
| continue; |
| if (const ConstantSDNode *IdxConst = |
| dyn_cast<ConstantSDNode>(U->getOperand(1))) { |
| if (IdxConst->getZExtValue() == 0) |
| E0.push_back(U); |
| else if (IdxConst->getZExtValue() == 1) |
| E1.push_back(U); |
| else |
| llvm_unreachable("Invalid vector index."); |
| } |
| } |
| |
| // There's no point scattering f16x2 if we only ever access one |
| // element of it. |
| if (E0.empty() || E1.empty()) |
| return false; |
| |
| unsigned Op = NVPTX::SplitF16x2; |
| // If the vector has been BITCAST'ed from i32, we can use original |
| // value directly and avoid register-to-register move. |
| SDValue Source = Vector; |
| if (Vector->getOpcode() == ISD::BITCAST) { |
| Op = NVPTX::SplitI32toF16x2; |
| Source = Vector->getOperand(0); |
| } |
| // Merge (f16 extractelt(V, 0), f16 extractelt(V,1)) |
| // into f16,f16 SplitF16x2(V) |
| SDNode *ScatterOp = |
| CurDAG->getMachineNode(Op, SDLoc(N), MVT::f16, MVT::f16, Source); |
| for (auto *Node : E0) |
| ReplaceUses(SDValue(Node, 0), SDValue(ScatterOp, 0)); |
| for (auto *Node : E1) |
| ReplaceUses(SDValue(Node, 0), SDValue(ScatterOp, 1)); |
| |
| return true; |
| } |
| |
| static unsigned int getCodeAddrSpace(MemSDNode *N) { |
| const Value *Src = N->getMemOperand()->getValue(); |
| |
| if (!Src) |
| return NVPTX::PTXLdStInstCode::GENERIC; |
| |
| if (auto *PT = dyn_cast<PointerType>(Src->getType())) { |
| switch (PT->getAddressSpace()) { |
| case llvm::ADDRESS_SPACE_LOCAL: return NVPTX::PTXLdStInstCode::LOCAL; |
| case llvm::ADDRESS_SPACE_GLOBAL: return NVPTX::PTXLdStInstCode::GLOBAL; |
| case llvm::ADDRESS_SPACE_SHARED: return NVPTX::PTXLdStInstCode::SHARED; |
| case llvm::ADDRESS_SPACE_GENERIC: return NVPTX::PTXLdStInstCode::GENERIC; |
| case llvm::ADDRESS_SPACE_PARAM: return NVPTX::PTXLdStInstCode::PARAM; |
| case llvm::ADDRESS_SPACE_CONST: return NVPTX::PTXLdStInstCode::CONSTANT; |
| default: break; |
| } |
| } |
| return NVPTX::PTXLdStInstCode::GENERIC; |
| } |
| |
| static bool canLowerToLDG(MemSDNode *N, const NVPTXSubtarget &Subtarget, |
| unsigned CodeAddrSpace, MachineFunction *F) { |
| // We use ldg (i.e. ld.global.nc) for invariant loads from the global address |
| // space. |
| // |
| // We have two ways of identifying invariant loads: Loads may be explicitly |
| // marked as invariant, or we may infer them to be invariant. |
| // |
| // We currently infer invariance for loads from |
| // - constant global variables, and |
| // - kernel function pointer params that are noalias (i.e. __restrict) and |
| // never written to. |
| // |
| // TODO: Perform a more powerful invariance analysis (ideally IPO, and ideally |
| // not during the SelectionDAG phase). |
| // |
| // TODO: Infer invariance only at -O2. We still want to use ldg at -O0 for |
| // explicitly invariant loads because these are how clang tells us to use ldg |
| // when the user uses a builtin. |
| if (!Subtarget.hasLDG() || CodeAddrSpace != NVPTX::PTXLdStInstCode::GLOBAL) |
| return false; |
| |
| if (N->isInvariant()) |
| return true; |
| |
| bool IsKernelFn = isKernelFunction(F->getFunction()); |
| |
| // We use GetUnderlyingObjects() here instead of GetUnderlyingObject() mainly |
| // because the former looks through phi nodes while the latter does not. We |
| // need to look through phi nodes to handle pointer induction variables. |
| SmallVector<Value *, 8> Objs; |
| GetUnderlyingObjects(const_cast<Value *>(N->getMemOperand()->getValue()), |
| Objs, F->getDataLayout()); |
| |
| return all_of(Objs, [&](Value *V) { |
| if (auto *A = dyn_cast<const Argument>(V)) |
| return IsKernelFn && A->onlyReadsMemory() && A->hasNoAliasAttr(); |
| if (auto *GV = dyn_cast<const GlobalVariable>(V)) |
| return GV->isConstant(); |
| return false; |
| }); |
| } |
| |
| bool NVPTXDAGToDAGISel::tryIntrinsicNoChain(SDNode *N) { |
| unsigned IID = cast<ConstantSDNode>(N->getOperand(0))->getZExtValue(); |
| switch (IID) { |
| default: |
| return false; |
| case Intrinsic::nvvm_texsurf_handle_internal: |
| SelectTexSurfHandle(N); |
| return true; |
| } |
| } |
| |
| void NVPTXDAGToDAGISel::SelectTexSurfHandle(SDNode *N) { |
| // Op 0 is the intrinsic ID |
| SDValue Wrapper = N->getOperand(1); |
| SDValue GlobalVal = Wrapper.getOperand(0); |
| ReplaceNode(N, CurDAG->getMachineNode(NVPTX::texsurf_handles, SDLoc(N), |
| MVT::i64, GlobalVal)); |
| } |
| |
| void NVPTXDAGToDAGISel::SelectAddrSpaceCast(SDNode *N) { |
| SDValue Src = N->getOperand(0); |
| AddrSpaceCastSDNode *CastN = cast<AddrSpaceCastSDNode>(N); |
| unsigned SrcAddrSpace = CastN->getSrcAddressSpace(); |
| unsigned DstAddrSpace = CastN->getDestAddressSpace(); |
| assert(SrcAddrSpace != DstAddrSpace && |
| "addrspacecast must be between different address spaces"); |
| |
| if (DstAddrSpace == ADDRESS_SPACE_GENERIC) { |
| // Specific to generic |
| unsigned Opc; |
| switch (SrcAddrSpace) { |
| default: report_fatal_error("Bad address space in addrspacecast"); |
| case ADDRESS_SPACE_GLOBAL: |
| Opc = TM.is64Bit() ? NVPTX::cvta_global_yes_64 : NVPTX::cvta_global_yes; |
| break; |
| case ADDRESS_SPACE_SHARED: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_shared_yes_6432 |
| : NVPTX::cvta_shared_yes_64) |
| : NVPTX::cvta_shared_yes; |
| break; |
| case ADDRESS_SPACE_CONST: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_const_yes_6432 |
| : NVPTX::cvta_const_yes_64) |
| : NVPTX::cvta_const_yes; |
| break; |
| case ADDRESS_SPACE_LOCAL: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_local_yes_6432 |
| : NVPTX::cvta_local_yes_64) |
| : NVPTX::cvta_local_yes; |
| break; |
| } |
| ReplaceNode(N, CurDAG->getMachineNode(Opc, SDLoc(N), N->getValueType(0), |
| Src)); |
| return; |
| } else { |
| // Generic to specific |
| if (SrcAddrSpace != 0) |
| report_fatal_error("Cannot cast between two non-generic address spaces"); |
| unsigned Opc; |
| switch (DstAddrSpace) { |
| default: report_fatal_error("Bad address space in addrspacecast"); |
| case ADDRESS_SPACE_GLOBAL: |
| Opc = TM.is64Bit() ? NVPTX::cvta_to_global_yes_64 |
| : NVPTX::cvta_to_global_yes; |
| break; |
| case ADDRESS_SPACE_SHARED: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_to_shared_yes_3264 |
| : NVPTX::cvta_to_shared_yes_64) |
| : NVPTX::cvta_to_shared_yes; |
| break; |
| case ADDRESS_SPACE_CONST: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_to_const_yes_3264 |
| : NVPTX::cvta_to_const_yes_64) |
| : NVPTX::cvta_to_const_yes; |
| break; |
| case ADDRESS_SPACE_LOCAL: |
| Opc = TM.is64Bit() ? (useShortPointers() ? NVPTX::cvta_to_local_yes_3264 |
| : NVPTX::cvta_to_local_yes_64) |
| : NVPTX::cvta_to_local_yes; |
| break; |
| case ADDRESS_SPACE_PARAM: |
| Opc = TM.is64Bit() ? NVPTX::nvvm_ptr_gen_to_param_64 |
| : NVPTX::nvvm_ptr_gen_to_param; |
| break; |
| } |
| ReplaceNode(N, CurDAG->getMachineNode(Opc, SDLoc(N), N->getValueType(0), |
| Src)); |
| return; |
| } |
| } |
| |
| // Helper function template to reduce amount of boilerplate code for |
| // opcode selection. |
| static Optional<unsigned> pickOpcodeForVT( |
| MVT::SimpleValueType VT, unsigned Opcode_i8, unsigned Opcode_i16, |
| unsigned Opcode_i32, Optional<unsigned> Opcode_i64, unsigned Opcode_f16, |
| unsigned Opcode_f16x2, unsigned Opcode_f32, Optional<unsigned> Opcode_f64) { |
| switch (VT) { |
| case MVT::i1: |
| case MVT::i8: |
| return Opcode_i8; |
| case MVT::i16: |
| return Opcode_i16; |
| case MVT::i32: |
| return Opcode_i32; |
| case MVT::i64: |
| return Opcode_i64; |
| case MVT::f16: |
| return Opcode_f16; |
| case MVT::v2f16: |
| return Opcode_f16x2; |
| case MVT::f32: |
| return Opcode_f32; |
| case MVT::f64: |
| return Opcode_f64; |
| default: |
| return None; |
| } |
| } |
| |
| bool NVPTXDAGToDAGISel::tryLoad(SDNode *N) { |
| SDLoc dl(N); |
| MemSDNode *LD = cast<MemSDNode>(N); |
| assert(LD->readMem() && "Expected load"); |
| LoadSDNode *PlainLoad = dyn_cast<LoadSDNode>(N); |
| EVT LoadedVT = LD->getMemoryVT(); |
| SDNode *NVPTXLD = nullptr; |
| |
| // do not support pre/post inc/dec |
| if (PlainLoad && PlainLoad->isIndexed()) |
| return false; |
| |
| if (!LoadedVT.isSimple()) |
| return false; |
| |
| AtomicOrdering Ordering = LD->getOrdering(); |
| // In order to lower atomic loads with stronger guarantees we would need to |
| // use load.acquire or insert fences. However these features were only added |
| // with PTX ISA 6.0 / sm_70. |
| // TODO: Check if we can actually use the new instructions and implement them. |
| if (isStrongerThanMonotonic(Ordering)) |
| return false; |
| |
| // Address Space Setting |
| unsigned int CodeAddrSpace = getCodeAddrSpace(LD); |
| if (canLowerToLDG(LD, *Subtarget, CodeAddrSpace, MF)) { |
| return tryLDGLDU(N); |
| } |
| |
| unsigned int PointerSize = |
| CurDAG->getDataLayout().getPointerSizeInBits(LD->getAddressSpace()); |
| |
| // Volatile Setting |
| // - .volatile is only available for .global and .shared |
| // - .volatile has the same memory synchronization semantics as .relaxed.sys |
| bool isVolatile = LD->isVolatile() || Ordering == AtomicOrdering::Monotonic; |
| if (CodeAddrSpace != NVPTX::PTXLdStInstCode::GLOBAL && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::SHARED && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::GENERIC) |
| isVolatile = false; |
| |
| // Type Setting: fromType + fromTypeWidth |
| // |
| // Sign : ISD::SEXTLOAD |
| // Unsign : ISD::ZEXTLOAD, ISD::NON_EXTLOAD or ISD::EXTLOAD and the |
| // type is integer |
| // Float : ISD::NON_EXTLOAD or ISD::EXTLOAD and the type is float |
| MVT SimpleVT = LoadedVT.getSimpleVT(); |
| MVT ScalarVT = SimpleVT.getScalarType(); |
| // Read at least 8 bits (predicates are stored as 8-bit values) |
| unsigned fromTypeWidth = std::max(8U, ScalarVT.getSizeInBits()); |
| unsigned int fromType; |
| |
| // Vector Setting |
| unsigned vecType = NVPTX::PTXLdStInstCode::Scalar; |
| if (SimpleVT.isVector()) { |
| assert(LoadedVT == MVT::v2f16 && "Unexpected vector type"); |
| // v2f16 is loaded using ld.b32 |
| fromTypeWidth = 32; |
| } |
| |
| if (PlainLoad && (PlainLoad->getExtensionType() == ISD::SEXTLOAD)) |
| fromType = NVPTX::PTXLdStInstCode::Signed; |
| else if (ScalarVT.isFloatingPoint()) |
| // f16 uses .b16 as its storage type. |
| fromType = ScalarVT.SimpleTy == MVT::f16 ? NVPTX::PTXLdStInstCode::Untyped |
| : NVPTX::PTXLdStInstCode::Float; |
| else |
| fromType = NVPTX::PTXLdStInstCode::Unsigned; |
| |
| // Create the machine instruction DAG |
| SDValue Chain = N->getOperand(0); |
| SDValue N1 = N->getOperand(1); |
| SDValue Addr; |
| SDValue Offset, Base; |
| Optional<unsigned> Opcode; |
| MVT::SimpleValueType TargetVT = LD->getSimpleValueType(0).SimpleTy; |
| |
| if (SelectDirectAddr(N1, Addr)) { |
| Opcode = pickOpcodeForVT( |
| TargetVT, NVPTX::LD_i8_avar, NVPTX::LD_i16_avar, NVPTX::LD_i32_avar, |
| NVPTX::LD_i64_avar, NVPTX::LD_f16_avar, NVPTX::LD_f16x2_avar, |
| NVPTX::LD_f32_avar, NVPTX::LD_f64_avar); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(isVolatile, dl), getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), getI32Imm(fromType, dl), |
| getI32Imm(fromTypeWidth, dl), Addr, Chain }; |
| NVPTXLD = CurDAG->getMachineNode(Opcode.getValue(), dl, TargetVT, |
| MVT::Other, Ops); |
| } else if (PointerSize == 64 ? SelectADDRsi64(N1.getNode(), N1, Base, Offset) |
| : SelectADDRsi(N1.getNode(), N1, Base, Offset)) { |
| Opcode = pickOpcodeForVT(TargetVT, NVPTX::LD_i8_asi, NVPTX::LD_i16_asi, |
| NVPTX::LD_i32_asi, NVPTX::LD_i64_asi, |
| NVPTX::LD_f16_asi, NVPTX::LD_f16x2_asi, |
| NVPTX::LD_f32_asi, NVPTX::LD_f64_asi); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(isVolatile, dl), getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), getI32Imm(fromType, dl), |
| getI32Imm(fromTypeWidth, dl), Base, Offset, Chain }; |
| NVPTXLD = CurDAG->getMachineNode(Opcode.getValue(), dl, TargetVT, |
| MVT::Other, Ops); |
| } else if (PointerSize == 64 ? SelectADDRri64(N1.getNode(), N1, Base, Offset) |
| : SelectADDRri(N1.getNode(), N1, Base, Offset)) { |
| if (PointerSize == 64) |
| Opcode = pickOpcodeForVT( |
| TargetVT, NVPTX::LD_i8_ari_64, NVPTX::LD_i16_ari_64, |
| NVPTX::LD_i32_ari_64, NVPTX::LD_i64_ari_64, NVPTX::LD_f16_ari_64, |
| NVPTX::LD_f16x2_ari_64, NVPTX::LD_f32_ari_64, NVPTX::LD_f64_ari_64); |
| else |
| Opcode = pickOpcodeForVT( |
| TargetVT, NVPTX::LD_i8_ari, NVPTX::LD_i16_ari, NVPTX::LD_i32_ari, |
| NVPTX::LD_i64_ari, NVPTX::LD_f16_ari, NVPTX::LD_f16x2_ari, |
| NVPTX::LD_f32_ari, NVPTX::LD_f64_ari); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(isVolatile, dl), getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), getI32Imm(fromType, dl), |
| getI32Imm(fromTypeWidth, dl), Base, Offset, Chain }; |
| NVPTXLD = CurDAG->getMachineNode(Opcode.getValue(), dl, TargetVT, |
| MVT::Other, Ops); |
| } else { |
| if (PointerSize == 64) |
| Opcode = pickOpcodeForVT( |
| TargetVT, NVPTX::LD_i8_areg_64, NVPTX::LD_i16_areg_64, |
| NVPTX::LD_i32_areg_64, NVPTX::LD_i64_areg_64, NVPTX::LD_f16_areg_64, |
| NVPTX::LD_f16x2_areg_64, NVPTX::LD_f32_areg_64, |
| NVPTX::LD_f64_areg_64); |
| else |
| Opcode = pickOpcodeForVT( |
| TargetVT, NVPTX::LD_i8_areg, NVPTX::LD_i16_areg, NVPTX::LD_i32_areg, |
| NVPTX::LD_i64_areg, NVPTX::LD_f16_areg, NVPTX::LD_f16x2_areg, |
| NVPTX::LD_f32_areg, NVPTX::LD_f64_areg); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(isVolatile, dl), getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), getI32Imm(fromType, dl), |
| getI32Imm(fromTypeWidth, dl), N1, Chain }; |
| NVPTXLD = CurDAG->getMachineNode(Opcode.getValue(), dl, TargetVT, |
| MVT::Other, Ops); |
| } |
| |
| if (!NVPTXLD) |
| return false; |
| |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(NVPTXLD)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| ReplaceNode(N, NVPTXLD); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryLoadVector(SDNode *N) { |
| |
| SDValue Chain = N->getOperand(0); |
| SDValue Op1 = N->getOperand(1); |
| SDValue Addr, Offset, Base; |
| Optional<unsigned> Opcode; |
| SDLoc DL(N); |
| SDNode *LD; |
| MemSDNode *MemSD = cast<MemSDNode>(N); |
| EVT LoadedVT = MemSD->getMemoryVT(); |
| |
| if (!LoadedVT.isSimple()) |
| return false; |
| |
| // Address Space Setting |
| unsigned int CodeAddrSpace = getCodeAddrSpace(MemSD); |
| if (canLowerToLDG(MemSD, *Subtarget, CodeAddrSpace, MF)) { |
| return tryLDGLDU(N); |
| } |
| |
| unsigned int PointerSize = |
| CurDAG->getDataLayout().getPointerSizeInBits(MemSD->getAddressSpace()); |
| |
| // Volatile Setting |
| // - .volatile is only availalble for .global and .shared |
| bool IsVolatile = MemSD->isVolatile(); |
| if (CodeAddrSpace != NVPTX::PTXLdStInstCode::GLOBAL && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::SHARED && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::GENERIC) |
| IsVolatile = false; |
| |
| // Vector Setting |
| MVT SimpleVT = LoadedVT.getSimpleVT(); |
| |
| // Type Setting: fromType + fromTypeWidth |
| // |
| // Sign : ISD::SEXTLOAD |
| // Unsign : ISD::ZEXTLOAD, ISD::NON_EXTLOAD or ISD::EXTLOAD and the |
| // type is integer |
| // Float : ISD::NON_EXTLOAD or ISD::EXTLOAD and the type is float |
| MVT ScalarVT = SimpleVT.getScalarType(); |
| // Read at least 8 bits (predicates are stored as 8-bit values) |
| unsigned FromTypeWidth = std::max(8U, ScalarVT.getSizeInBits()); |
| unsigned int FromType; |
| // The last operand holds the original LoadSDNode::getExtensionType() value |
| unsigned ExtensionType = cast<ConstantSDNode>( |
| N->getOperand(N->getNumOperands() - 1))->getZExtValue(); |
| if (ExtensionType == ISD::SEXTLOAD) |
| FromType = NVPTX::PTXLdStInstCode::Signed; |
| else if (ScalarVT.isFloatingPoint()) |
| FromType = ScalarVT.SimpleTy == MVT::f16 ? NVPTX::PTXLdStInstCode::Untyped |
| : NVPTX::PTXLdStInstCode::Float; |
| else |
| FromType = NVPTX::PTXLdStInstCode::Unsigned; |
| |
| unsigned VecType; |
| |
| switch (N->getOpcode()) { |
| case NVPTXISD::LoadV2: |
| VecType = NVPTX::PTXLdStInstCode::V2; |
| break; |
| case NVPTXISD::LoadV4: |
| VecType = NVPTX::PTXLdStInstCode::V4; |
| break; |
| default: |
| return false; |
| } |
| |
| EVT EltVT = N->getValueType(0); |
| |
| // v8f16 is a special case. PTX doesn't have ld.v8.f16 |
| // instruction. Instead, we split the vector into v2f16 chunks and |
| // load them with ld.v4.b32. |
| if (EltVT == MVT::v2f16) { |
| assert(N->getOpcode() == NVPTXISD::LoadV4 && "Unexpected load opcode."); |
| EltVT = MVT::i32; |
| FromType = NVPTX::PTXLdStInstCode::Untyped; |
| FromTypeWidth = 32; |
| } |
| |
| if (SelectDirectAddr(Op1, Addr)) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::LDV_i8_v2_avar, NVPTX::LDV_i16_v2_avar, |
| NVPTX::LDV_i32_v2_avar, NVPTX::LDV_i64_v2_avar, |
| NVPTX::LDV_f16_v2_avar, NVPTX::LDV_f16x2_v2_avar, |
| NVPTX::LDV_f32_v2_avar, NVPTX::LDV_f64_v2_avar); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_avar, |
| NVPTX::LDV_i16_v4_avar, NVPTX::LDV_i32_v4_avar, None, |
| NVPTX::LDV_f16_v4_avar, NVPTX::LDV_f16x2_v4_avar, |
| NVPTX::LDV_f32_v4_avar, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(IsVolatile, DL), getI32Imm(CodeAddrSpace, DL), |
| getI32Imm(VecType, DL), getI32Imm(FromType, DL), |
| getI32Imm(FromTypeWidth, DL), Addr, Chain }; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, N->getVTList(), Ops); |
| } else if (PointerSize == 64 |
| ? SelectADDRsi64(Op1.getNode(), Op1, Base, Offset) |
| : SelectADDRsi(Op1.getNode(), Op1, Base, Offset)) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::LDV_i8_v2_asi, NVPTX::LDV_i16_v2_asi, |
| NVPTX::LDV_i32_v2_asi, NVPTX::LDV_i64_v2_asi, |
| NVPTX::LDV_f16_v2_asi, NVPTX::LDV_f16x2_v2_asi, |
| NVPTX::LDV_f32_v2_asi, NVPTX::LDV_f64_v2_asi); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_asi, |
| NVPTX::LDV_i16_v4_asi, NVPTX::LDV_i32_v4_asi, None, |
| NVPTX::LDV_f16_v4_asi, NVPTX::LDV_f16x2_v4_asi, |
| NVPTX::LDV_f32_v4_asi, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(IsVolatile, DL), getI32Imm(CodeAddrSpace, DL), |
| getI32Imm(VecType, DL), getI32Imm(FromType, DL), |
| getI32Imm(FromTypeWidth, DL), Base, Offset, Chain }; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, N->getVTList(), Ops); |
| } else if (PointerSize == 64 |
| ? SelectADDRri64(Op1.getNode(), Op1, Base, Offset) |
| : SelectADDRri(Op1.getNode(), Op1, Base, Offset)) { |
| if (PointerSize == 64) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v2_ari_64, |
| NVPTX::LDV_i16_v2_ari_64, NVPTX::LDV_i32_v2_ari_64, |
| NVPTX::LDV_i64_v2_ari_64, NVPTX::LDV_f16_v2_ari_64, |
| NVPTX::LDV_f16x2_v2_ari_64, NVPTX::LDV_f32_v2_ari_64, |
| NVPTX::LDV_f64_v2_ari_64); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_ari_64, |
| NVPTX::LDV_i16_v4_ari_64, NVPTX::LDV_i32_v4_ari_64, None, |
| NVPTX::LDV_f16_v4_ari_64, NVPTX::LDV_f16x2_v4_ari_64, |
| NVPTX::LDV_f32_v4_ari_64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::LDV_i8_v2_ari, NVPTX::LDV_i16_v2_ari, |
| NVPTX::LDV_i32_v2_ari, NVPTX::LDV_i64_v2_ari, |
| NVPTX::LDV_f16_v2_ari, NVPTX::LDV_f16x2_v2_ari, |
| NVPTX::LDV_f32_v2_ari, NVPTX::LDV_f64_v2_ari); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_ari, |
| NVPTX::LDV_i16_v4_ari, NVPTX::LDV_i32_v4_ari, None, |
| NVPTX::LDV_f16_v4_ari, NVPTX::LDV_f16x2_v4_ari, |
| NVPTX::LDV_f32_v4_ari, None); |
| break; |
| } |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(IsVolatile, DL), getI32Imm(CodeAddrSpace, DL), |
| getI32Imm(VecType, DL), getI32Imm(FromType, DL), |
| getI32Imm(FromTypeWidth, DL), Base, Offset, Chain }; |
| |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, N->getVTList(), Ops); |
| } else { |
| if (PointerSize == 64) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v2_areg_64, |
| NVPTX::LDV_i16_v2_areg_64, NVPTX::LDV_i32_v2_areg_64, |
| NVPTX::LDV_i64_v2_areg_64, NVPTX::LDV_f16_v2_areg_64, |
| NVPTX::LDV_f16x2_v2_areg_64, NVPTX::LDV_f32_v2_areg_64, |
| NVPTX::LDV_f64_v2_areg_64); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_areg_64, |
| NVPTX::LDV_i16_v4_areg_64, NVPTX::LDV_i32_v4_areg_64, None, |
| NVPTX::LDV_f16_v4_areg_64, NVPTX::LDV_f16x2_v4_areg_64, |
| NVPTX::LDV_f32_v4_areg_64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadV2: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v2_areg, |
| NVPTX::LDV_i16_v2_areg, NVPTX::LDV_i32_v2_areg, |
| NVPTX::LDV_i64_v2_areg, NVPTX::LDV_f16_v2_areg, |
| NVPTX::LDV_f16x2_v2_areg, NVPTX::LDV_f32_v2_areg, |
| NVPTX::LDV_f64_v2_areg); |
| break; |
| case NVPTXISD::LoadV4: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::LDV_i8_v4_areg, |
| NVPTX::LDV_i16_v4_areg, NVPTX::LDV_i32_v4_areg, None, |
| NVPTX::LDV_f16_v4_areg, NVPTX::LDV_f16x2_v4_areg, |
| NVPTX::LDV_f32_v4_areg, None); |
| break; |
| } |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { getI32Imm(IsVolatile, DL), getI32Imm(CodeAddrSpace, DL), |
| getI32Imm(VecType, DL), getI32Imm(FromType, DL), |
| getI32Imm(FromTypeWidth, DL), Op1, Chain }; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, N->getVTList(), Ops); |
| } |
| |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(LD)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| ReplaceNode(N, LD); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryLDGLDU(SDNode *N) { |
| |
| SDValue Chain = N->getOperand(0); |
| SDValue Op1; |
| MemSDNode *Mem; |
| bool IsLDG = true; |
| |
| // If this is an LDG intrinsic, the address is the third operand. If its an |
| // LDG/LDU SD node (from custom vector handling), then its the second operand |
| if (N->getOpcode() == ISD::INTRINSIC_W_CHAIN) { |
| Op1 = N->getOperand(2); |
| Mem = cast<MemIntrinsicSDNode>(N); |
| unsigned IID = cast<ConstantSDNode>(N->getOperand(1))->getZExtValue(); |
| switch (IID) { |
| default: |
| return false; |
| case Intrinsic::nvvm_ldg_global_f: |
| case Intrinsic::nvvm_ldg_global_i: |
| case Intrinsic::nvvm_ldg_global_p: |
| IsLDG = true; |
| break; |
| case Intrinsic::nvvm_ldu_global_f: |
| case Intrinsic::nvvm_ldu_global_i: |
| case Intrinsic::nvvm_ldu_global_p: |
| IsLDG = false; |
| break; |
| } |
| } else { |
| Op1 = N->getOperand(1); |
| Mem = cast<MemSDNode>(N); |
| } |
| |
| Optional<unsigned> Opcode; |
| SDLoc DL(N); |
| SDNode *LD; |
| SDValue Base, Offset, Addr; |
| |
| EVT EltVT = Mem->getMemoryVT(); |
| unsigned NumElts = 1; |
| if (EltVT.isVector()) { |
| NumElts = EltVT.getVectorNumElements(); |
| EltVT = EltVT.getVectorElementType(); |
| // vectors of f16 are loaded/stored as multiples of v2f16 elements. |
| if (EltVT == MVT::f16 && N->getValueType(0) == MVT::v2f16) { |
| assert(NumElts % 2 == 0 && "Vector must have even number of elements"); |
| EltVT = MVT::v2f16; |
| NumElts /= 2; |
| } |
| } |
| |
| // Build the "promoted" result VTList for the load. If we are really loading |
| // i8s, then the return type will be promoted to i16 since we do not expose |
| // 8-bit registers in NVPTX. |
| EVT NodeVT = (EltVT == MVT::i8) ? MVT::i16 : EltVT; |
| SmallVector<EVT, 5> InstVTs; |
| for (unsigned i = 0; i != NumElts; ++i) { |
| InstVTs.push_back(NodeVT); |
| } |
| InstVTs.push_back(MVT::Other); |
| SDVTList InstVTList = CurDAG->getVTList(InstVTs); |
| |
| if (SelectDirectAddr(Op1, Addr)) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case ISD::LOAD: |
| case ISD::INTRINSIC_W_CHAIN: |
| if (IsLDG) |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_GLOBAL_i8avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_i16avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_i32avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_i64avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16x2avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_f32avar, |
| NVPTX::INT_PTX_LDG_GLOBAL_f64avar); |
| else |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_GLOBAL_i8avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_i16avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_i32avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_i64avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16x2avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_f32avar, |
| NVPTX::INT_PTX_LDU_GLOBAL_f64avar); |
| break; |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LDGV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v2i8_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2i16_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2i32_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2i64_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2f16_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2f16x2_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2f32_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v2f64_ELE_avar); |
| break; |
| case NVPTXISD::LDUV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v2i8_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2i16_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2i32_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2i64_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2f16_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2f16x2_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2f32_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v2f64_ELE_avar); |
| break; |
| case NVPTXISD::LoadV4: |
| case NVPTXISD::LDGV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v4i8_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v4i16_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v4i32_ELE_avar, None, |
| NVPTX::INT_PTX_LDG_G_v4f16_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v4f16x2_ELE_avar, |
| NVPTX::INT_PTX_LDG_G_v4f32_ELE_avar, None); |
| break; |
| case NVPTXISD::LDUV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v4i8_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v4i16_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v4i32_ELE_avar, None, |
| NVPTX::INT_PTX_LDU_G_v4f16_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v4f16x2_ELE_avar, |
| NVPTX::INT_PTX_LDU_G_v4f32_ELE_avar, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { Addr, Chain }; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, InstVTList, Ops); |
| } else if (TM.is64Bit() ? SelectADDRri64(Op1.getNode(), Op1, Base, Offset) |
| : SelectADDRri(Op1.getNode(), Op1, Base, Offset)) { |
| if (TM.is64Bit()) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case ISD::LOAD: |
| case ISD::INTRINSIC_W_CHAIN: |
| if (IsLDG) |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_GLOBAL_i8ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i16ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i32ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i64ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16x2ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f32ari64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f64ari64); |
| else |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_GLOBAL_i8ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i16ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i32ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i64ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16x2ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f32ari64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f64ari64); |
| break; |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LDGV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v2i8_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2i16_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2i32_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2i64_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2f16_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2f16x2_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2f32_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v2f64_ELE_ari64); |
| break; |
| case NVPTXISD::LDUV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v2i8_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2i16_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2i32_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2i64_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2f16_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2f16x2_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2f32_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v2f64_ELE_ari64); |
| break; |
| case NVPTXISD::LoadV4: |
| case NVPTXISD::LDGV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v4i8_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v4i16_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v4i32_ELE_ari64, None, |
| NVPTX::INT_PTX_LDG_G_v4f16_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v4f16x2_ELE_ari64, |
| NVPTX::INT_PTX_LDG_G_v4f32_ELE_ari64, None); |
| break; |
| case NVPTXISD::LDUV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v4i8_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v4i16_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v4i32_ELE_ari64, None, |
| NVPTX::INT_PTX_LDU_G_v4f16_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v4f16x2_ELE_ari64, |
| NVPTX::INT_PTX_LDU_G_v4f32_ELE_ari64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case ISD::LOAD: |
| case ISD::INTRINSIC_W_CHAIN: |
| if (IsLDG) |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_GLOBAL_i8ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_i16ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_i32ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_i64ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16x2ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_f32ari, |
| NVPTX::INT_PTX_LDG_GLOBAL_f64ari); |
| else |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_GLOBAL_i8ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_i16ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_i32ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_i64ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16x2ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_f32ari, |
| NVPTX::INT_PTX_LDU_GLOBAL_f64ari); |
| break; |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LDGV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v2i8_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2i16_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2i32_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2i64_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2f16_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2f16x2_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2f32_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v2f64_ELE_ari32); |
| break; |
| case NVPTXISD::LDUV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v2i8_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2i16_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2i32_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2i64_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2f16_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2f16x2_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2f32_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v2f64_ELE_ari32); |
| break; |
| case NVPTXISD::LoadV4: |
| case NVPTXISD::LDGV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v4i8_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v4i16_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v4i32_ELE_ari32, None, |
| NVPTX::INT_PTX_LDG_G_v4f16_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v4f16x2_ELE_ari32, |
| NVPTX::INT_PTX_LDG_G_v4f32_ELE_ari32, None); |
| break; |
| case NVPTXISD::LDUV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v4i8_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v4i16_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v4i32_ELE_ari32, None, |
| NVPTX::INT_PTX_LDU_G_v4f16_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v4f16x2_ELE_ari32, |
| NVPTX::INT_PTX_LDU_G_v4f32_ELE_ari32, None); |
| break; |
| } |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = {Base, Offset, Chain}; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, InstVTList, Ops); |
| } else { |
| if (TM.is64Bit()) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case ISD::LOAD: |
| case ISD::INTRINSIC_W_CHAIN: |
| if (IsLDG) |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_GLOBAL_i8areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i16areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i32areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_i64areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16x2areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f32areg64, |
| NVPTX::INT_PTX_LDG_GLOBAL_f64areg64); |
| else |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_GLOBAL_i8areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i16areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i32areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_i64areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16x2areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f32areg64, |
| NVPTX::INT_PTX_LDU_GLOBAL_f64areg64); |
| break; |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LDGV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v2i8_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2i16_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2i32_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2i64_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2f16_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2f16x2_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2f32_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v2f64_ELE_areg64); |
| break; |
| case NVPTXISD::LDUV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v2i8_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2i16_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2i32_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2i64_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2f16_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2f16x2_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2f32_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v2f64_ELE_areg64); |
| break; |
| case NVPTXISD::LoadV4: |
| case NVPTXISD::LDGV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v4i8_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v4i16_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v4i32_ELE_areg64, None, |
| NVPTX::INT_PTX_LDG_G_v4f16_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v4f16x2_ELE_areg64, |
| NVPTX::INT_PTX_LDG_G_v4f32_ELE_areg64, None); |
| break; |
| case NVPTXISD::LDUV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v4i8_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v4i16_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v4i32_ELE_areg64, None, |
| NVPTX::INT_PTX_LDU_G_v4f16_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v4f16x2_ELE_areg64, |
| NVPTX::INT_PTX_LDU_G_v4f32_ELE_areg64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case ISD::LOAD: |
| case ISD::INTRINSIC_W_CHAIN: |
| if (IsLDG) |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_GLOBAL_i8areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_i16areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_i32areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_i64areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_f16x2areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_f32areg, |
| NVPTX::INT_PTX_LDG_GLOBAL_f64areg); |
| else |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_GLOBAL_i8areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_i16areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_i32areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_i64areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_f16x2areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_f32areg, |
| NVPTX::INT_PTX_LDU_GLOBAL_f64areg); |
| break; |
| case NVPTXISD::LoadV2: |
| case NVPTXISD::LDGV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v2i8_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2i16_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2i32_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2i64_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2f16_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2f16x2_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2f32_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v2f64_ELE_areg32); |
| break; |
| case NVPTXISD::LDUV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v2i8_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2i16_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2i32_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2i64_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2f16_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2f16x2_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2f32_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v2f64_ELE_areg32); |
| break; |
| case NVPTXISD::LoadV4: |
| case NVPTXISD::LDGV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDG_G_v4i8_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v4i16_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v4i32_ELE_areg32, None, |
| NVPTX::INT_PTX_LDG_G_v4f16_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v4f16x2_ELE_areg32, |
| NVPTX::INT_PTX_LDG_G_v4f32_ELE_areg32, None); |
| break; |
| case NVPTXISD::LDUV4: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::INT_PTX_LDU_G_v4i8_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v4i16_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v4i32_ELE_areg32, None, |
| NVPTX::INT_PTX_LDU_G_v4f16_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v4f16x2_ELE_areg32, |
| NVPTX::INT_PTX_LDU_G_v4f32_ELE_areg32, None); |
| break; |
| } |
| } |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = { Op1, Chain }; |
| LD = CurDAG->getMachineNode(Opcode.getValue(), DL, InstVTList, Ops); |
| } |
| |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = Mem->getMemOperand(); |
| cast<MachineSDNode>(LD)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| // For automatic generation of LDG (through SelectLoad[Vector], not the |
| // intrinsics), we may have an extending load like: |
| // |
| // i32,ch = load<LD1[%data1(addrspace=1)], zext from i8> t0, t7, undef:i64 |
| // |
| // In this case, the matching logic above will select a load for the original |
| // memory type (in this case, i8) and our types will not match (the node needs |
| // to return an i32 in this case). Our LDG/LDU nodes do not support the |
| // concept of sign-/zero-extension, so emulate it here by adding an explicit |
| // CVT instruction. Ptxas should clean up any redundancies here. |
| |
| EVT OrigType = N->getValueType(0); |
| LoadSDNode *LdNode = dyn_cast<LoadSDNode>(N); |
| |
| if (OrigType != EltVT && LdNode) { |
| // We have an extending-load. The instruction we selected operates on the |
| // smaller type, but the SDNode we are replacing has the larger type. We |
| // need to emit a CVT to make the types match. |
| bool IsSigned = LdNode->getExtensionType() == ISD::SEXTLOAD; |
| unsigned CvtOpc = GetConvertOpcode(OrigType.getSimpleVT(), |
| EltVT.getSimpleVT(), IsSigned); |
| |
| // For each output value, apply the manual sign/zero-extension and make sure |
| // all users of the load go through that CVT. |
| for (unsigned i = 0; i != NumElts; ++i) { |
| SDValue Res(LD, i); |
| SDValue OrigVal(N, i); |
| |
| SDNode *CvtNode = |
| CurDAG->getMachineNode(CvtOpc, DL, OrigType, Res, |
| CurDAG->getTargetConstant(NVPTX::PTXCvtMode::NONE, |
| DL, MVT::i32)); |
| ReplaceUses(OrigVal, SDValue(CvtNode, 0)); |
| } |
| } |
| |
| ReplaceNode(N, LD); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryStore(SDNode *N) { |
| SDLoc dl(N); |
| MemSDNode *ST = cast<MemSDNode>(N); |
| assert(ST->writeMem() && "Expected store"); |
| StoreSDNode *PlainStore = dyn_cast<StoreSDNode>(N); |
| AtomicSDNode *AtomicStore = dyn_cast<AtomicSDNode>(N); |
| assert((PlainStore || AtomicStore) && "Expected store"); |
| EVT StoreVT = ST->getMemoryVT(); |
| SDNode *NVPTXST = nullptr; |
| |
| // do not support pre/post inc/dec |
| if (PlainStore && PlainStore->isIndexed()) |
| return false; |
| |
| if (!StoreVT.isSimple()) |
| return false; |
| |
| AtomicOrdering Ordering = ST->getOrdering(); |
| // In order to lower atomic loads with stronger guarantees we would need to |
| // use store.release or insert fences. However these features were only added |
| // with PTX ISA 6.0 / sm_70. |
| // TODO: Check if we can actually use the new instructions and implement them. |
| if (isStrongerThanMonotonic(Ordering)) |
| return false; |
| |
| // Address Space Setting |
| unsigned int CodeAddrSpace = getCodeAddrSpace(ST); |
| unsigned int PointerSize = |
| CurDAG->getDataLayout().getPointerSizeInBits(ST->getAddressSpace()); |
| |
| // Volatile Setting |
| // - .volatile is only available for .global and .shared |
| // - .volatile has the same memory synchronization semantics as .relaxed.sys |
| bool isVolatile = ST->isVolatile() || Ordering == AtomicOrdering::Monotonic; |
| if (CodeAddrSpace != NVPTX::PTXLdStInstCode::GLOBAL && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::SHARED && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::GENERIC) |
| isVolatile = false; |
| |
| // Vector Setting |
| MVT SimpleVT = StoreVT.getSimpleVT(); |
| unsigned vecType = NVPTX::PTXLdStInstCode::Scalar; |
| |
| // Type Setting: toType + toTypeWidth |
| // - for integer type, always use 'u' |
| // |
| MVT ScalarVT = SimpleVT.getScalarType(); |
| unsigned toTypeWidth = ScalarVT.getSizeInBits(); |
| if (SimpleVT.isVector()) { |
| assert(StoreVT == MVT::v2f16 && "Unexpected vector type"); |
| // v2f16 is stored using st.b32 |
| toTypeWidth = 32; |
| } |
| |
| unsigned int toType; |
| if (ScalarVT.isFloatingPoint()) |
| // f16 uses .b16 as its storage type. |
| toType = ScalarVT.SimpleTy == MVT::f16 ? NVPTX::PTXLdStInstCode::Untyped |
| : NVPTX::PTXLdStInstCode::Float; |
| else |
| toType = NVPTX::PTXLdStInstCode::Unsigned; |
| |
| // Create the machine instruction DAG |
| SDValue Chain = ST->getChain(); |
| SDValue Value = PlainStore ? PlainStore->getValue() : AtomicStore->getVal(); |
| SDValue BasePtr = ST->getBasePtr(); |
| SDValue Addr; |
| SDValue Offset, Base; |
| Optional<unsigned> Opcode; |
| MVT::SimpleValueType SourceVT = |
| Value.getNode()->getSimpleValueType(0).SimpleTy; |
| |
| if (SelectDirectAddr(BasePtr, Addr)) { |
| Opcode = pickOpcodeForVT(SourceVT, NVPTX::ST_i8_avar, NVPTX::ST_i16_avar, |
| NVPTX::ST_i32_avar, NVPTX::ST_i64_avar, |
| NVPTX::ST_f16_avar, NVPTX::ST_f16x2_avar, |
| NVPTX::ST_f32_avar, NVPTX::ST_f64_avar); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = {Value, |
| getI32Imm(isVolatile, dl), |
| getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), |
| getI32Imm(toType, dl), |
| getI32Imm(toTypeWidth, dl), |
| Addr, |
| Chain}; |
| NVPTXST = CurDAG->getMachineNode(Opcode.getValue(), dl, MVT::Other, Ops); |
| } else if (PointerSize == 64 |
| ? SelectADDRsi64(BasePtr.getNode(), BasePtr, Base, Offset) |
| : SelectADDRsi(BasePtr.getNode(), BasePtr, Base, Offset)) { |
| Opcode = pickOpcodeForVT(SourceVT, NVPTX::ST_i8_asi, NVPTX::ST_i16_asi, |
| NVPTX::ST_i32_asi, NVPTX::ST_i64_asi, |
| NVPTX::ST_f16_asi, NVPTX::ST_f16x2_asi, |
| NVPTX::ST_f32_asi, NVPTX::ST_f64_asi); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = {Value, |
| getI32Imm(isVolatile, dl), |
| getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), |
| getI32Imm(toType, dl), |
| getI32Imm(toTypeWidth, dl), |
| Base, |
| Offset, |
| Chain}; |
| NVPTXST = CurDAG->getMachineNode(Opcode.getValue(), dl, MVT::Other, Ops); |
| } else if (PointerSize == 64 |
| ? SelectADDRri64(BasePtr.getNode(), BasePtr, Base, Offset) |
| : SelectADDRri(BasePtr.getNode(), BasePtr, Base, Offset)) { |
| if (PointerSize == 64) |
| Opcode = pickOpcodeForVT( |
| SourceVT, NVPTX::ST_i8_ari_64, NVPTX::ST_i16_ari_64, |
| NVPTX::ST_i32_ari_64, NVPTX::ST_i64_ari_64, NVPTX::ST_f16_ari_64, |
| NVPTX::ST_f16x2_ari_64, NVPTX::ST_f32_ari_64, NVPTX::ST_f64_ari_64); |
| else |
| Opcode = pickOpcodeForVT(SourceVT, NVPTX::ST_i8_ari, NVPTX::ST_i16_ari, |
| NVPTX::ST_i32_ari, NVPTX::ST_i64_ari, |
| NVPTX::ST_f16_ari, NVPTX::ST_f16x2_ari, |
| NVPTX::ST_f32_ari, NVPTX::ST_f64_ari); |
| if (!Opcode) |
| return false; |
| |
| SDValue Ops[] = {Value, |
| getI32Imm(isVolatile, dl), |
| getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), |
| getI32Imm(toType, dl), |
| getI32Imm(toTypeWidth, dl), |
| Base, |
| Offset, |
| Chain}; |
| NVPTXST = CurDAG->getMachineNode(Opcode.getValue(), dl, MVT::Other, Ops); |
| } else { |
| if (PointerSize == 64) |
| Opcode = |
| pickOpcodeForVT(SourceVT, NVPTX::ST_i8_areg_64, NVPTX::ST_i16_areg_64, |
| NVPTX::ST_i32_areg_64, NVPTX::ST_i64_areg_64, |
| NVPTX::ST_f16_areg_64, NVPTX::ST_f16x2_areg_64, |
| NVPTX::ST_f32_areg_64, NVPTX::ST_f64_areg_64); |
| else |
| Opcode = pickOpcodeForVT(SourceVT, NVPTX::ST_i8_areg, NVPTX::ST_i16_areg, |
| NVPTX::ST_i32_areg, NVPTX::ST_i64_areg, |
| NVPTX::ST_f16_areg, NVPTX::ST_f16x2_areg, |
| NVPTX::ST_f32_areg, NVPTX::ST_f64_areg); |
| if (!Opcode) |
| return false; |
| SDValue Ops[] = {Value, |
| getI32Imm(isVolatile, dl), |
| getI32Imm(CodeAddrSpace, dl), |
| getI32Imm(vecType, dl), |
| getI32Imm(toType, dl), |
| getI32Imm(toTypeWidth, dl), |
| BasePtr, |
| Chain}; |
| NVPTXST = CurDAG->getMachineNode(Opcode.getValue(), dl, MVT::Other, Ops); |
| } |
| |
| if (!NVPTXST) |
| return false; |
| |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(NVPTXST)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| ReplaceNode(N, NVPTXST); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryStoreVector(SDNode *N) { |
| SDValue Chain = N->getOperand(0); |
| SDValue Op1 = N->getOperand(1); |
| SDValue Addr, Offset, Base; |
| Optional<unsigned> Opcode; |
| SDLoc DL(N); |
| SDNode *ST; |
| EVT EltVT = Op1.getValueType(); |
| MemSDNode *MemSD = cast<MemSDNode>(N); |
| EVT StoreVT = MemSD->getMemoryVT(); |
| |
| // Address Space Setting |
| unsigned CodeAddrSpace = getCodeAddrSpace(MemSD); |
| if (CodeAddrSpace == NVPTX::PTXLdStInstCode::CONSTANT) { |
| report_fatal_error("Cannot store to pointer that points to constant " |
| "memory space"); |
| } |
| unsigned int PointerSize = |
| CurDAG->getDataLayout().getPointerSizeInBits(MemSD->getAddressSpace()); |
| |
| // Volatile Setting |
| // - .volatile is only availalble for .global and .shared |
| bool IsVolatile = MemSD->isVolatile(); |
| if (CodeAddrSpace != NVPTX::PTXLdStInstCode::GLOBAL && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::SHARED && |
| CodeAddrSpace != NVPTX::PTXLdStInstCode::GENERIC) |
| IsVolatile = false; |
| |
| // Type Setting: toType + toTypeWidth |
| // - for integer type, always use 'u' |
| assert(StoreVT.isSimple() && "Store value is not simple"); |
| MVT ScalarVT = StoreVT.getSimpleVT().getScalarType(); |
| unsigned ToTypeWidth = ScalarVT.getSizeInBits(); |
| unsigned ToType; |
| if (ScalarVT.isFloatingPoint()) |
| ToType = ScalarVT.SimpleTy == MVT::f16 ? NVPTX::PTXLdStInstCode::Untyped |
| : NVPTX::PTXLdStInstCode::Float; |
| else |
| ToType = NVPTX::PTXLdStInstCode::Unsigned; |
| |
| SmallVector<SDValue, 12> StOps; |
| SDValue N2; |
| unsigned VecType; |
| |
| switch (N->getOpcode()) { |
| case NVPTXISD::StoreV2: |
| VecType = NVPTX::PTXLdStInstCode::V2; |
| StOps.push_back(N->getOperand(1)); |
| StOps.push_back(N->getOperand(2)); |
| N2 = N->getOperand(3); |
| break; |
| case NVPTXISD::StoreV4: |
| VecType = NVPTX::PTXLdStInstCode::V4; |
| StOps.push_back(N->getOperand(1)); |
| StOps.push_back(N->getOperand(2)); |
| StOps.push_back(N->getOperand(3)); |
| StOps.push_back(N->getOperand(4)); |
| N2 = N->getOperand(5); |
| break; |
| default: |
| return false; |
| } |
| |
| // v8f16 is a special case. PTX doesn't have st.v8.f16 |
| // instruction. Instead, we split the vector into v2f16 chunks and |
| // store them with st.v4.b32. |
| if (EltVT == MVT::v2f16) { |
| assert(N->getOpcode() == NVPTXISD::StoreV4 && "Unexpected load opcode."); |
| EltVT = MVT::i32; |
| ToType = NVPTX::PTXLdStInstCode::Untyped; |
| ToTypeWidth = 32; |
| } |
| |
| StOps.push_back(getI32Imm(IsVolatile, DL)); |
| StOps.push_back(getI32Imm(CodeAddrSpace, DL)); |
| StOps.push_back(getI32Imm(VecType, DL)); |
| StOps.push_back(getI32Imm(ToType, DL)); |
| StOps.push_back(getI32Imm(ToTypeWidth, DL)); |
| |
| if (SelectDirectAddr(N2, Addr)) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::STV_i8_v2_avar, NVPTX::STV_i16_v2_avar, |
| NVPTX::STV_i32_v2_avar, NVPTX::STV_i64_v2_avar, |
| NVPTX::STV_f16_v2_avar, NVPTX::STV_f16x2_v2_avar, |
| NVPTX::STV_f32_v2_avar, NVPTX::STV_f64_v2_avar); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_avar, |
| NVPTX::STV_i16_v4_avar, NVPTX::STV_i32_v4_avar, None, |
| NVPTX::STV_f16_v4_avar, NVPTX::STV_f16x2_v4_avar, |
| NVPTX::STV_f32_v4_avar, None); |
| break; |
| } |
| StOps.push_back(Addr); |
| } else if (PointerSize == 64 ? SelectADDRsi64(N2.getNode(), N2, Base, Offset) |
| : SelectADDRsi(N2.getNode(), N2, Base, Offset)) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::STV_i8_v2_asi, NVPTX::STV_i16_v2_asi, |
| NVPTX::STV_i32_v2_asi, NVPTX::STV_i64_v2_asi, |
| NVPTX::STV_f16_v2_asi, NVPTX::STV_f16x2_v2_asi, |
| NVPTX::STV_f32_v2_asi, NVPTX::STV_f64_v2_asi); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_asi, |
| NVPTX::STV_i16_v4_asi, NVPTX::STV_i32_v4_asi, None, |
| NVPTX::STV_f16_v4_asi, NVPTX::STV_f16x2_v4_asi, |
| NVPTX::STV_f32_v4_asi, None); |
| break; |
| } |
| StOps.push_back(Base); |
| StOps.push_back(Offset); |
| } else if (PointerSize == 64 ? SelectADDRri64(N2.getNode(), N2, Base, Offset) |
| : SelectADDRri(N2.getNode(), N2, Base, Offset)) { |
| if (PointerSize == 64) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v2_ari_64, |
| NVPTX::STV_i16_v2_ari_64, NVPTX::STV_i32_v2_ari_64, |
| NVPTX::STV_i64_v2_ari_64, NVPTX::STV_f16_v2_ari_64, |
| NVPTX::STV_f16x2_v2_ari_64, NVPTX::STV_f32_v2_ari_64, |
| NVPTX::STV_f64_v2_ari_64); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_ari_64, |
| NVPTX::STV_i16_v4_ari_64, NVPTX::STV_i32_v4_ari_64, None, |
| NVPTX::STV_f16_v4_ari_64, NVPTX::STV_f16x2_v4_ari_64, |
| NVPTX::STV_f32_v4_ari_64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, |
| NVPTX::STV_i8_v2_ari, NVPTX::STV_i16_v2_ari, |
| NVPTX::STV_i32_v2_ari, NVPTX::STV_i64_v2_ari, |
| NVPTX::STV_f16_v2_ari, NVPTX::STV_f16x2_v2_ari, |
| NVPTX::STV_f32_v2_ari, NVPTX::STV_f64_v2_ari); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_ari, |
| NVPTX::STV_i16_v4_ari, NVPTX::STV_i32_v4_ari, None, |
| NVPTX::STV_f16_v4_ari, NVPTX::STV_f16x2_v4_ari, |
| NVPTX::STV_f32_v4_ari, None); |
| break; |
| } |
| } |
| StOps.push_back(Base); |
| StOps.push_back(Offset); |
| } else { |
| if (PointerSize == 64) { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v2_areg_64, |
| NVPTX::STV_i16_v2_areg_64, NVPTX::STV_i32_v2_areg_64, |
| NVPTX::STV_i64_v2_areg_64, NVPTX::STV_f16_v2_areg_64, |
| NVPTX::STV_f16x2_v2_areg_64, NVPTX::STV_f32_v2_areg_64, |
| NVPTX::STV_f64_v2_areg_64); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = pickOpcodeForVT( |
| EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_areg_64, |
| NVPTX::STV_i16_v4_areg_64, NVPTX::STV_i32_v4_areg_64, None, |
| NVPTX::STV_f16_v4_areg_64, NVPTX::STV_f16x2_v4_areg_64, |
| NVPTX::STV_f32_v4_areg_64, None); |
| break; |
| } |
| } else { |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreV2: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v2_areg, |
| NVPTX::STV_i16_v2_areg, NVPTX::STV_i32_v2_areg, |
| NVPTX::STV_i64_v2_areg, NVPTX::STV_f16_v2_areg, |
| NVPTX::STV_f16x2_v2_areg, NVPTX::STV_f32_v2_areg, |
| NVPTX::STV_f64_v2_areg); |
| break; |
| case NVPTXISD::StoreV4: |
| Opcode = |
| pickOpcodeForVT(EltVT.getSimpleVT().SimpleTy, NVPTX::STV_i8_v4_areg, |
| NVPTX::STV_i16_v4_areg, NVPTX::STV_i32_v4_areg, None, |
| NVPTX::STV_f16_v4_areg, NVPTX::STV_f16x2_v4_areg, |
| NVPTX::STV_f32_v4_areg, None); |
| break; |
| } |
| } |
| StOps.push_back(N2); |
| } |
| |
| if (!Opcode) |
| return false; |
| |
| StOps.push_back(Chain); |
| |
| ST = CurDAG->getMachineNode(Opcode.getValue(), DL, MVT::Other, StOps); |
| |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(ST)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| ReplaceNode(N, ST); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryLoadParam(SDNode *Node) { |
| SDValue Chain = Node->getOperand(0); |
| SDValue Offset = Node->getOperand(2); |
| SDValue Flag = Node->getOperand(3); |
| SDLoc DL(Node); |
| MemSDNode *Mem = cast<MemSDNode>(Node); |
| |
| unsigned VecSize; |
| switch (Node->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::LoadParam: |
| VecSize = 1; |
| break; |
| case NVPTXISD::LoadParamV2: |
| VecSize = 2; |
| break; |
| case NVPTXISD::LoadParamV4: |
| VecSize = 4; |
| break; |
| } |
| |
| EVT EltVT = Node->getValueType(0); |
| EVT MemVT = Mem->getMemoryVT(); |
| |
| Optional<unsigned> Opcode; |
| |
| switch (VecSize) { |
| default: |
| return false; |
| case 1: |
| Opcode = pickOpcodeForVT(MemVT.getSimpleVT().SimpleTy, |
| NVPTX::LoadParamMemI8, NVPTX::LoadParamMemI16, |
| NVPTX::LoadParamMemI32, NVPTX::LoadParamMemI64, |
| NVPTX::LoadParamMemF16, NVPTX::LoadParamMemF16x2, |
| NVPTX::LoadParamMemF32, NVPTX::LoadParamMemF64); |
| break; |
| case 2: |
| Opcode = |
| pickOpcodeForVT(MemVT.getSimpleVT().SimpleTy, NVPTX::LoadParamMemV2I8, |
| NVPTX::LoadParamMemV2I16, NVPTX::LoadParamMemV2I32, |
| NVPTX::LoadParamMemV2I64, NVPTX::LoadParamMemV2F16, |
| NVPTX::LoadParamMemV2F16x2, NVPTX::LoadParamMemV2F32, |
| NVPTX::LoadParamMemV2F64); |
| break; |
| case 4: |
| Opcode = pickOpcodeForVT( |
| MemVT.getSimpleVT().SimpleTy, NVPTX::LoadParamMemV4I8, |
| NVPTX::LoadParamMemV4I16, NVPTX::LoadParamMemV4I32, None, |
| NVPTX::LoadParamMemV4F16, NVPTX::LoadParamMemV4F16x2, |
| NVPTX::LoadParamMemV4F32, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| |
| SDVTList VTs; |
| if (VecSize == 1) { |
| VTs = CurDAG->getVTList(EltVT, MVT::Other, MVT::Glue); |
| } else if (VecSize == 2) { |
| VTs = CurDAG->getVTList(EltVT, EltVT, MVT::Other, MVT::Glue); |
| } else { |
| EVT EVTs[] = { EltVT, EltVT, EltVT, EltVT, MVT::Other, MVT::Glue }; |
| VTs = CurDAG->getVTList(EVTs); |
| } |
| |
| unsigned OffsetVal = cast<ConstantSDNode>(Offset)->getZExtValue(); |
| |
| SmallVector<SDValue, 2> Ops; |
| Ops.push_back(CurDAG->getTargetConstant(OffsetVal, DL, MVT::i32)); |
| Ops.push_back(Chain); |
| Ops.push_back(Flag); |
| |
| ReplaceNode(Node, CurDAG->getMachineNode(Opcode.getValue(), DL, VTs, Ops)); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryStoreRetval(SDNode *N) { |
| SDLoc DL(N); |
| SDValue Chain = N->getOperand(0); |
| SDValue Offset = N->getOperand(1); |
| unsigned OffsetVal = cast<ConstantSDNode>(Offset)->getZExtValue(); |
| MemSDNode *Mem = cast<MemSDNode>(N); |
| |
| // How many elements do we have? |
| unsigned NumElts = 1; |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreRetval: |
| NumElts = 1; |
| break; |
| case NVPTXISD::StoreRetvalV2: |
| NumElts = 2; |
| break; |
| case NVPTXISD::StoreRetvalV4: |
| NumElts = 4; |
| break; |
| } |
| |
| // Build vector of operands |
| SmallVector<SDValue, 6> Ops; |
| for (unsigned i = 0; i < NumElts; ++i) |
| Ops.push_back(N->getOperand(i + 2)); |
| Ops.push_back(CurDAG->getTargetConstant(OffsetVal, DL, MVT::i32)); |
| Ops.push_back(Chain); |
| |
| // Determine target opcode |
| // If we have an i1, use an 8-bit store. The lowering code in |
| // NVPTXISelLowering will have already emitted an upcast. |
| Optional<unsigned> Opcode = 0; |
| switch (NumElts) { |
| default: |
| return false; |
| case 1: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreRetvalI8, NVPTX::StoreRetvalI16, |
| NVPTX::StoreRetvalI32, NVPTX::StoreRetvalI64, |
| NVPTX::StoreRetvalF16, NVPTX::StoreRetvalF16x2, |
| NVPTX::StoreRetvalF32, NVPTX::StoreRetvalF64); |
| break; |
| case 2: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreRetvalV2I8, NVPTX::StoreRetvalV2I16, |
| NVPTX::StoreRetvalV2I32, NVPTX::StoreRetvalV2I64, |
| NVPTX::StoreRetvalV2F16, NVPTX::StoreRetvalV2F16x2, |
| NVPTX::StoreRetvalV2F32, NVPTX::StoreRetvalV2F64); |
| break; |
| case 4: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreRetvalV4I8, NVPTX::StoreRetvalV4I16, |
| NVPTX::StoreRetvalV4I32, None, |
| NVPTX::StoreRetvalV4F16, NVPTX::StoreRetvalV4F16x2, |
| NVPTX::StoreRetvalV4F32, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| |
| SDNode *Ret = CurDAG->getMachineNode(Opcode.getValue(), DL, MVT::Other, Ops); |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(Ret)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| ReplaceNode(N, Ret); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryStoreParam(SDNode *N) { |
| SDLoc DL(N); |
| SDValue Chain = N->getOperand(0); |
| SDValue Param = N->getOperand(1); |
| unsigned ParamVal = cast<ConstantSDNode>(Param)->getZExtValue(); |
| SDValue Offset = N->getOperand(2); |
| unsigned OffsetVal = cast<ConstantSDNode>(Offset)->getZExtValue(); |
| MemSDNode *Mem = cast<MemSDNode>(N); |
| SDValue Flag = N->getOperand(N->getNumOperands() - 1); |
| |
| // How many elements do we have? |
| unsigned NumElts = 1; |
| switch (N->getOpcode()) { |
| default: |
| return false; |
| case NVPTXISD::StoreParamU32: |
| case NVPTXISD::StoreParamS32: |
| case NVPTXISD::StoreParam: |
| NumElts = 1; |
| break; |
| case NVPTXISD::StoreParamV2: |
| NumElts = 2; |
| break; |
| case NVPTXISD::StoreParamV4: |
| NumElts = 4; |
| break; |
| } |
| |
| // Build vector of operands |
| SmallVector<SDValue, 8> Ops; |
| for (unsigned i = 0; i < NumElts; ++i) |
| Ops.push_back(N->getOperand(i + 3)); |
| Ops.push_back(CurDAG->getTargetConstant(ParamVal, DL, MVT::i32)); |
| Ops.push_back(CurDAG->getTargetConstant(OffsetVal, DL, MVT::i32)); |
| Ops.push_back(Chain); |
| Ops.push_back(Flag); |
| |
| // Determine target opcode |
| // If we have an i1, use an 8-bit store. The lowering code in |
| // NVPTXISelLowering will have already emitted an upcast. |
| Optional<unsigned> Opcode = 0; |
| switch (N->getOpcode()) { |
| default: |
| switch (NumElts) { |
| default: |
| return false; |
| case 1: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreParamI8, NVPTX::StoreParamI16, |
| NVPTX::StoreParamI32, NVPTX::StoreParamI64, |
| NVPTX::StoreParamF16, NVPTX::StoreParamF16x2, |
| NVPTX::StoreParamF32, NVPTX::StoreParamF64); |
| break; |
| case 2: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreParamV2I8, NVPTX::StoreParamV2I16, |
| NVPTX::StoreParamV2I32, NVPTX::StoreParamV2I64, |
| NVPTX::StoreParamV2F16, NVPTX::StoreParamV2F16x2, |
| NVPTX::StoreParamV2F32, NVPTX::StoreParamV2F64); |
| break; |
| case 4: |
| Opcode = pickOpcodeForVT(Mem->getMemoryVT().getSimpleVT().SimpleTy, |
| NVPTX::StoreParamV4I8, NVPTX::StoreParamV4I16, |
| NVPTX::StoreParamV4I32, None, |
| NVPTX::StoreParamV4F16, NVPTX::StoreParamV4F16x2, |
| NVPTX::StoreParamV4F32, None); |
| break; |
| } |
| if (!Opcode) |
| return false; |
| break; |
| // Special case: if we have a sign-extend/zero-extend node, insert the |
| // conversion instruction first, and use that as the value operand to |
| // the selected StoreParam node. |
| case NVPTXISD::StoreParamU32: { |
| Opcode = NVPTX::StoreParamI32; |
| SDValue CvtNone = CurDAG->getTargetConstant(NVPTX::PTXCvtMode::NONE, DL, |
| MVT::i32); |
| SDNode *Cvt = CurDAG->getMachineNode(NVPTX::CVT_u32_u16, DL, |
| MVT::i32, Ops[0], CvtNone); |
| Ops[0] = SDValue(Cvt, 0); |
| break; |
| } |
| case NVPTXISD::StoreParamS32: { |
| Opcode = NVPTX::StoreParamI32; |
| SDValue CvtNone = CurDAG->getTargetConstant(NVPTX::PTXCvtMode::NONE, DL, |
| MVT::i32); |
| SDNode *Cvt = CurDAG->getMachineNode(NVPTX::CVT_s32_s16, DL, |
| MVT::i32, Ops[0], CvtNone); |
| Ops[0] = SDValue(Cvt, 0); |
| break; |
| } |
| } |
| |
| SDVTList RetVTs = CurDAG->getVTList(MVT::Other, MVT::Glue); |
| SDNode *Ret = |
| CurDAG->getMachineNode(Opcode.getValue(), DL, RetVTs, Ops); |
| MachineSDNode::mmo_iterator MemRefs0 = MF->allocateMemRefsArray(1); |
| MemRefs0[0] = cast<MemSDNode>(N)->getMemOperand(); |
| cast<MachineSDNode>(Ret)->setMemRefs(MemRefs0, MemRefs0 + 1); |
| |
| ReplaceNode(N, Ret); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::tryTextureIntrinsic(SDNode *N) { |
| unsigned Opc = 0; |
| |
| switch (N->getOpcode()) { |
| default: return false; |
| case NVPTXISD::Tex1DFloatS32: |
| Opc = NVPTX::TEX_1D_F32_S32; |
| break; |
| case NVPTXISD::Tex1DFloatFloat: |
| Opc = NVPTX::TEX_1D_F32_F32; |
| break; |
| case NVPTXISD::Tex1DFloatFloatLevel: |
| Opc = NVPTX::TEX_1D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DFloatFloatGrad: |
| Opc = NVPTX::TEX_1D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex1DS32S32: |
| Opc = NVPTX::TEX_1D_S32_S32; |
| break; |
| case NVPTXISD::Tex1DS32Float: |
| Opc = NVPTX::TEX_1D_S32_F32; |
| break; |
| case NVPTXISD::Tex1DS32FloatLevel: |
| Opc = NVPTX::TEX_1D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DS32FloatGrad: |
| Opc = NVPTX::TEX_1D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex1DU32S32: |
| Opc = NVPTX::TEX_1D_U32_S32; |
| break; |
| case NVPTXISD::Tex1DU32Float: |
| Opc = NVPTX::TEX_1D_U32_F32; |
| break; |
| case NVPTXISD::Tex1DU32FloatLevel: |
| Opc = NVPTX::TEX_1D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DU32FloatGrad: |
| Opc = NVPTX::TEX_1D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex1DArrayFloatS32: |
| Opc = NVPTX::TEX_1D_ARRAY_F32_S32; |
| break; |
| case NVPTXISD::Tex1DArrayFloatFloat: |
| Opc = NVPTX::TEX_1D_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::Tex1DArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_1D_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DArrayFloatFloatGrad: |
| Opc = NVPTX::TEX_1D_ARRAY_F32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex1DArrayS32S32: |
| Opc = NVPTX::TEX_1D_ARRAY_S32_S32; |
| break; |
| case NVPTXISD::Tex1DArrayS32Float: |
| Opc = NVPTX::TEX_1D_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::Tex1DArrayS32FloatLevel: |
| Opc = NVPTX::TEX_1D_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DArrayS32FloatGrad: |
| Opc = NVPTX::TEX_1D_ARRAY_S32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex1DArrayU32S32: |
| Opc = NVPTX::TEX_1D_ARRAY_U32_S32; |
| break; |
| case NVPTXISD::Tex1DArrayU32Float: |
| Opc = NVPTX::TEX_1D_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::Tex1DArrayU32FloatLevel: |
| Opc = NVPTX::TEX_1D_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex1DArrayU32FloatGrad: |
| Opc = NVPTX::TEX_1D_ARRAY_U32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DFloatS32: |
| Opc = NVPTX::TEX_2D_F32_S32; |
| break; |
| case NVPTXISD::Tex2DFloatFloat: |
| Opc = NVPTX::TEX_2D_F32_F32; |
| break; |
| case NVPTXISD::Tex2DFloatFloatLevel: |
| Opc = NVPTX::TEX_2D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DFloatFloatGrad: |
| Opc = NVPTX::TEX_2D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DS32S32: |
| Opc = NVPTX::TEX_2D_S32_S32; |
| break; |
| case NVPTXISD::Tex2DS32Float: |
| Opc = NVPTX::TEX_2D_S32_F32; |
| break; |
| case NVPTXISD::Tex2DS32FloatLevel: |
| Opc = NVPTX::TEX_2D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DS32FloatGrad: |
| Opc = NVPTX::TEX_2D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DU32S32: |
| Opc = NVPTX::TEX_2D_U32_S32; |
| break; |
| case NVPTXISD::Tex2DU32Float: |
| Opc = NVPTX::TEX_2D_U32_F32; |
| break; |
| case NVPTXISD::Tex2DU32FloatLevel: |
| Opc = NVPTX::TEX_2D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DU32FloatGrad: |
| Opc = NVPTX::TEX_2D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DArrayFloatS32: |
| Opc = NVPTX::TEX_2D_ARRAY_F32_S32; |
| break; |
| case NVPTXISD::Tex2DArrayFloatFloat: |
| Opc = NVPTX::TEX_2D_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::Tex2DArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_2D_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DArrayFloatFloatGrad: |
| Opc = NVPTX::TEX_2D_ARRAY_F32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DArrayS32S32: |
| Opc = NVPTX::TEX_2D_ARRAY_S32_S32; |
| break; |
| case NVPTXISD::Tex2DArrayS32Float: |
| Opc = NVPTX::TEX_2D_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::Tex2DArrayS32FloatLevel: |
| Opc = NVPTX::TEX_2D_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DArrayS32FloatGrad: |
| Opc = NVPTX::TEX_2D_ARRAY_S32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex2DArrayU32S32: |
| Opc = NVPTX::TEX_2D_ARRAY_U32_S32; |
| break; |
| case NVPTXISD::Tex2DArrayU32Float: |
| Opc = NVPTX::TEX_2D_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::Tex2DArrayU32FloatLevel: |
| Opc = NVPTX::TEX_2D_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex2DArrayU32FloatGrad: |
| Opc = NVPTX::TEX_2D_ARRAY_U32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex3DFloatS32: |
| Opc = NVPTX::TEX_3D_F32_S32; |
| break; |
| case NVPTXISD::Tex3DFloatFloat: |
| Opc = NVPTX::TEX_3D_F32_F32; |
| break; |
| case NVPTXISD::Tex3DFloatFloatLevel: |
| Opc = NVPTX::TEX_3D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex3DFloatFloatGrad: |
| Opc = NVPTX::TEX_3D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex3DS32S32: |
| Opc = NVPTX::TEX_3D_S32_S32; |
| break; |
| case NVPTXISD::Tex3DS32Float: |
| Opc = NVPTX::TEX_3D_S32_F32; |
| break; |
| case NVPTXISD::Tex3DS32FloatLevel: |
| Opc = NVPTX::TEX_3D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex3DS32FloatGrad: |
| Opc = NVPTX::TEX_3D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::Tex3DU32S32: |
| Opc = NVPTX::TEX_3D_U32_S32; |
| break; |
| case NVPTXISD::Tex3DU32Float: |
| Opc = NVPTX::TEX_3D_U32_F32; |
| break; |
| case NVPTXISD::Tex3DU32FloatLevel: |
| Opc = NVPTX::TEX_3D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tex3DU32FloatGrad: |
| Opc = NVPTX::TEX_3D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexCubeFloatFloat: |
| Opc = NVPTX::TEX_CUBE_F32_F32; |
| break; |
| case NVPTXISD::TexCubeFloatFloatLevel: |
| Opc = NVPTX::TEX_CUBE_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexCubeS32Float: |
| Opc = NVPTX::TEX_CUBE_S32_F32; |
| break; |
| case NVPTXISD::TexCubeS32FloatLevel: |
| Opc = NVPTX::TEX_CUBE_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexCubeU32Float: |
| Opc = NVPTX::TEX_CUBE_U32_F32; |
| break; |
| case NVPTXISD::TexCubeU32FloatLevel: |
| Opc = NVPTX::TEX_CUBE_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexCubeArrayFloatFloat: |
| Opc = NVPTX::TEX_CUBE_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::TexCubeArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_CUBE_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexCubeArrayS32Float: |
| Opc = NVPTX::TEX_CUBE_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::TexCubeArrayS32FloatLevel: |
| Opc = NVPTX::TEX_CUBE_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexCubeArrayU32Float: |
| Opc = NVPTX::TEX_CUBE_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::TexCubeArrayU32FloatLevel: |
| Opc = NVPTX::TEX_CUBE_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tld4R2DFloatFloat: |
| Opc = NVPTX::TLD4_R_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4G2DFloatFloat: |
| Opc = NVPTX::TLD4_G_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4B2DFloatFloat: |
| Opc = NVPTX::TLD4_B_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4A2DFloatFloat: |
| Opc = NVPTX::TLD4_A_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4R2DS64Float: |
| Opc = NVPTX::TLD4_R_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4G2DS64Float: |
| Opc = NVPTX::TLD4_G_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4B2DS64Float: |
| Opc = NVPTX::TLD4_B_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4A2DS64Float: |
| Opc = NVPTX::TLD4_A_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4R2DU64Float: |
| Opc = NVPTX::TLD4_R_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4G2DU64Float: |
| Opc = NVPTX::TLD4_G_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4B2DU64Float: |
| Opc = NVPTX::TLD4_B_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4A2DU64Float: |
| Opc = NVPTX::TLD4_A_2D_U32_F32; |
| break; |
| case NVPTXISD::TexUnified1DFloatS32: |
| Opc = NVPTX::TEX_UNIFIED_1D_F32_S32; |
| break; |
| case NVPTXISD::TexUnified1DFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_1D_F32_F32; |
| break; |
| case NVPTXISD::TexUnified1DFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DFloatFloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified1DS32S32: |
| Opc = NVPTX::TEX_UNIFIED_1D_S32_S32; |
| break; |
| case NVPTXISD::TexUnified1DS32Float: |
| Opc = NVPTX::TEX_UNIFIED_1D_S32_F32; |
| break; |
| case NVPTXISD::TexUnified1DS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DS32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified1DU32S32: |
| Opc = NVPTX::TEX_UNIFIED_1D_U32_S32; |
| break; |
| case NVPTXISD::TexUnified1DU32Float: |
| Opc = NVPTX::TEX_UNIFIED_1D_U32_F32; |
| break; |
| case NVPTXISD::TexUnified1DU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DU32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified1DArrayFloatS32: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_F32_S32; |
| break; |
| case NVPTXISD::TexUnified1DArrayFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::TexUnified1DArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DArrayFloatFloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_F32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified1DArrayS32S32: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_S32_S32; |
| break; |
| case NVPTXISD::TexUnified1DArrayS32Float: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::TexUnified1DArrayS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DArrayS32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_S32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified1DArrayU32S32: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_U32_S32; |
| break; |
| case NVPTXISD::TexUnified1DArrayU32Float: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::TexUnified1DArrayU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified1DArrayU32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_1D_ARRAY_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DFloatS32: |
| Opc = NVPTX::TEX_UNIFIED_2D_F32_S32; |
| break; |
| case NVPTXISD::TexUnified2DFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_2D_F32_F32; |
| break; |
| case NVPTXISD::TexUnified2DFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DFloatFloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DS32S32: |
| Opc = NVPTX::TEX_UNIFIED_2D_S32_S32; |
| break; |
| case NVPTXISD::TexUnified2DS32Float: |
| Opc = NVPTX::TEX_UNIFIED_2D_S32_F32; |
| break; |
| case NVPTXISD::TexUnified2DS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DS32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DU32S32: |
| Opc = NVPTX::TEX_UNIFIED_2D_U32_S32; |
| break; |
| case NVPTXISD::TexUnified2DU32Float: |
| Opc = NVPTX::TEX_UNIFIED_2D_U32_F32; |
| break; |
| case NVPTXISD::TexUnified2DU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DU32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DArrayFloatS32: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_F32_S32; |
| break; |
| case NVPTXISD::TexUnified2DArrayFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::TexUnified2DArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DArrayFloatFloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_F32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DArrayS32S32: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_S32_S32; |
| break; |
| case NVPTXISD::TexUnified2DArrayS32Float: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::TexUnified2DArrayS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DArrayS32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_S32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified2DArrayU32S32: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_U32_S32; |
| break; |
| case NVPTXISD::TexUnified2DArrayU32Float: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::TexUnified2DArrayU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified2DArrayU32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_2D_ARRAY_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified3DFloatS32: |
| Opc = NVPTX::TEX_UNIFIED_3D_F32_S32; |
| break; |
| case NVPTXISD::TexUnified3DFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_3D_F32_F32; |
| break; |
| case NVPTXISD::TexUnified3DFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_3D_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified3DFloatFloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_3D_F32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified3DS32S32: |
| Opc = NVPTX::TEX_UNIFIED_3D_S32_S32; |
| break; |
| case NVPTXISD::TexUnified3DS32Float: |
| Opc = NVPTX::TEX_UNIFIED_3D_S32_F32; |
| break; |
| case NVPTXISD::TexUnified3DS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_3D_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified3DS32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_3D_S32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnified3DU32S32: |
| Opc = NVPTX::TEX_UNIFIED_3D_U32_S32; |
| break; |
| case NVPTXISD::TexUnified3DU32Float: |
| Opc = NVPTX::TEX_UNIFIED_3D_U32_F32; |
| break; |
| case NVPTXISD::TexUnified3DU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_3D_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnified3DU32FloatGrad: |
| Opc = NVPTX::TEX_UNIFIED_3D_U32_F32_GRAD; |
| break; |
| case NVPTXISD::TexUnifiedCubeFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_F32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnifiedCubeS32Float: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_S32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnifiedCubeU32Float: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_U32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayFloatFloat: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_F32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayFloatFloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_F32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayS32Float: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_S32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayS32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_S32_F32_LEVEL; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayU32Float: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_U32_F32; |
| break; |
| case NVPTXISD::TexUnifiedCubeArrayU32FloatLevel: |
| Opc = NVPTX::TEX_UNIFIED_CUBE_ARRAY_U32_F32_LEVEL; |
| break; |
| case NVPTXISD::Tld4UnifiedR2DFloatFloat: |
| Opc = NVPTX::TLD4_UNIFIED_R_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedG2DFloatFloat: |
| Opc = NVPTX::TLD4_UNIFIED_G_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedB2DFloatFloat: |
| Opc = NVPTX::TLD4_UNIFIED_B_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedA2DFloatFloat: |
| Opc = NVPTX::TLD4_UNIFIED_A_2D_F32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedR2DS64Float: |
| Opc = NVPTX::TLD4_UNIFIED_R_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedG2DS64Float: |
| Opc = NVPTX::TLD4_UNIFIED_G_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedB2DS64Float: |
| Opc = NVPTX::TLD4_UNIFIED_B_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedA2DS64Float: |
| Opc = NVPTX::TLD4_UNIFIED_A_2D_S32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedR2DU64Float: |
| Opc = NVPTX::TLD4_UNIFIED_R_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedG2DU64Float: |
| Opc = NVPTX::TLD4_UNIFIED_G_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedB2DU64Float: |
| Opc = NVPTX::TLD4_UNIFIED_B_2D_U32_F32; |
| break; |
| case NVPTXISD::Tld4UnifiedA2DU64Float: |
| Opc = NVPTX::TLD4_UNIFIED_A_2D_U32_F32; |
| break; |
| } |
| |
| // Copy over operands |
| SmallVector<SDValue, 8> Ops(N->op_begin() + 1, N->op_end()); |
| Ops.push_back(N->getOperand(0)); // Move chain to the back. |
| |
| ReplaceNode(N, CurDAG->getMachineNode(Opc, SDLoc(N), N->getVTList(), Ops)); |
| return true; |
| } |
| |
| bool NVPTXDAGToDAGISel::trySurfaceIntrinsic(SDNode *N) { |
| unsigned Opc = 0; |
| switch (N->getOpcode()) { |
| default: return false; |
| case NVPTXISD::Suld1DI8Clamp: |
| Opc = NVPTX::SULD_1D_I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DI16Clamp: |
| Opc = NVPTX::SULD_1D_I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DI32Clamp: |
| Opc = NVPTX::SULD_1D_I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DI64Clamp: |
| Opc = NVPTX::SULD_1D_I64_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV2I8Clamp: |
| Opc = NVPTX::SULD_1D_V2I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV2I16Clamp: |
| Opc = NVPTX::SULD_1D_V2I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV2I32Clamp: |
| Opc = NVPTX::SULD_1D_V2I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV2I64Clamp: |
| Opc = NVPTX::SULD_1D_V2I64_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV4I8Clamp: |
| Opc = NVPTX::SULD_1D_V4I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV4I16Clamp: |
| Opc = NVPTX::SULD_1D_V4I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DV4I32Clamp: |
| Opc = NVPTX::SULD_1D_V4I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayI8Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayI16Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayI32Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayI64Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_I64_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I8Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I16Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I32Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I64Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I64_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I8Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I8_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I16Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I16_CLAMP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I32Clamp: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DI8Clamp: |
| Opc = NVPTX::SULD_2D_I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DI16Clamp: |
| Opc = NVPTX::SULD_2D_I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DI32Clamp: |
| Opc = NVPTX::SULD_2D_I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DI64Clamp: |
| Opc = NVPTX::SULD_2D_I64_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV2I8Clamp: |
| Opc = NVPTX::SULD_2D_V2I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV2I16Clamp: |
| Opc = NVPTX::SULD_2D_V2I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV2I32Clamp: |
| Opc = NVPTX::SULD_2D_V2I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV2I64Clamp: |
| Opc = NVPTX::SULD_2D_V2I64_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV4I8Clamp: |
| Opc = NVPTX::SULD_2D_V4I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV4I16Clamp: |
| Opc = NVPTX::SULD_2D_V4I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DV4I32Clamp: |
| Opc = NVPTX::SULD_2D_V4I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayI8Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayI16Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayI32Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayI64Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_I64_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I8Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I16Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I32Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I32_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I64Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I64_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I8Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I8_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I16Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I16_CLAMP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I32Clamp: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I32_CLAMP; |
| break; |
| case NVPTXISD::Suld3DI8Clamp: |
| Opc = NVPTX::SULD_3D_I8_CLAMP; |
| break; |
| case NVPTXISD::Suld3DI16Clamp: |
| Opc = NVPTX::SULD_3D_I16_CLAMP; |
| break; |
| case NVPTXISD::Suld3DI32Clamp: |
| Opc = NVPTX::SULD_3D_I32_CLAMP; |
| break; |
| case NVPTXISD::Suld3DI64Clamp: |
| Opc = NVPTX::SULD_3D_I64_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV2I8Clamp: |
| Opc = NVPTX::SULD_3D_V2I8_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV2I16Clamp: |
| Opc = NVPTX::SULD_3D_V2I16_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV2I32Clamp: |
| Opc = NVPTX::SULD_3D_V2I32_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV2I64Clamp: |
| Opc = NVPTX::SULD_3D_V2I64_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV4I8Clamp: |
| Opc = NVPTX::SULD_3D_V4I8_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV4I16Clamp: |
| Opc = NVPTX::SULD_3D_V4I16_CLAMP; |
| break; |
| case NVPTXISD::Suld3DV4I32Clamp: |
| Opc = NVPTX::SULD_3D_V4I32_CLAMP; |
| break; |
| case NVPTXISD::Suld1DI8Trap: |
| Opc = NVPTX::SULD_1D_I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DI16Trap: |
| Opc = NVPTX::SULD_1D_I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DI32Trap: |
| Opc = NVPTX::SULD_1D_I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DI64Trap: |
| Opc = NVPTX::SULD_1D_I64_TRAP; |
| break; |
| case NVPTXISD::Suld1DV2I8Trap: |
| Opc = NVPTX::SULD_1D_V2I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DV2I16Trap: |
| Opc = NVPTX::SULD_1D_V2I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DV2I32Trap: |
| Opc = NVPTX::SULD_1D_V2I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DV2I64Trap: |
| Opc = NVPTX::SULD_1D_V2I64_TRAP; |
| break; |
| case NVPTXISD::Suld1DV4I8Trap: |
| Opc = NVPTX::SULD_1D_V4I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DV4I16Trap: |
| Opc = NVPTX::SULD_1D_V4I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DV4I32Trap: |
| Opc = NVPTX::SULD_1D_V4I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayI8Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayI16Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayI32Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayI64Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_I64_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I8Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I16Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I32Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV2I64Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I64_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I8Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I8_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I16Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I16_TRAP; |
| break; |
| case NVPTXISD::Suld1DArrayV4I32Trap: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DI8Trap: |
| Opc = NVPTX::SULD_2D_I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DI16Trap: |
| Opc = NVPTX::SULD_2D_I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DI32Trap: |
| Opc = NVPTX::SULD_2D_I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DI64Trap: |
| Opc = NVPTX::SULD_2D_I64_TRAP; |
| break; |
| case NVPTXISD::Suld2DV2I8Trap: |
| Opc = NVPTX::SULD_2D_V2I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DV2I16Trap: |
| Opc = NVPTX::SULD_2D_V2I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DV2I32Trap: |
| Opc = NVPTX::SULD_2D_V2I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DV2I64Trap: |
| Opc = NVPTX::SULD_2D_V2I64_TRAP; |
| break; |
| case NVPTXISD::Suld2DV4I8Trap: |
| Opc = NVPTX::SULD_2D_V4I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DV4I16Trap: |
| Opc = NVPTX::SULD_2D_V4I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DV4I32Trap: |
| Opc = NVPTX::SULD_2D_V4I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayI8Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayI16Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayI32Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayI64Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_I64_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I8Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I16Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I32Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I32_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV2I64Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I64_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I8Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I8_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I16Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I16_TRAP; |
| break; |
| case NVPTXISD::Suld2DArrayV4I32Trap: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I32_TRAP; |
| break; |
| case NVPTXISD::Suld3DI8Trap: |
| Opc = NVPTX::SULD_3D_I8_TRAP; |
| break; |
| case NVPTXISD::Suld3DI16Trap: |
| Opc = NVPTX::SULD_3D_I16_TRAP; |
| break; |
| case NVPTXISD::Suld3DI32Trap: |
| Opc = NVPTX::SULD_3D_I32_TRAP; |
| break; |
| case NVPTXISD::Suld3DI64Trap: |
| Opc = NVPTX::SULD_3D_I64_TRAP; |
| break; |
| case NVPTXISD::Suld3DV2I8Trap: |
| Opc = NVPTX::SULD_3D_V2I8_TRAP; |
| break; |
| case NVPTXISD::Suld3DV2I16Trap: |
| Opc = NVPTX::SULD_3D_V2I16_TRAP; |
| break; |
| case NVPTXISD::Suld3DV2I32Trap: |
| Opc = NVPTX::SULD_3D_V2I32_TRAP; |
| break; |
| case NVPTXISD::Suld3DV2I64Trap: |
| Opc = NVPTX::SULD_3D_V2I64_TRAP; |
| break; |
| case NVPTXISD::Suld3DV4I8Trap: |
| Opc = NVPTX::SULD_3D_V4I8_TRAP; |
| break; |
| case NVPTXISD::Suld3DV4I16Trap: |
| Opc = NVPTX::SULD_3D_V4I16_TRAP; |
| break; |
| case NVPTXISD::Suld3DV4I32Trap: |
| Opc = NVPTX::SULD_3D_V4I32_TRAP; |
| break; |
| case NVPTXISD::Suld1DI8Zero: |
| Opc = NVPTX::SULD_1D_I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DI16Zero: |
| Opc = NVPTX::SULD_1D_I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DI32Zero: |
| Opc = NVPTX::SULD_1D_I32_ZERO; |
| break; |
| case NVPTXISD::Suld1DI64Zero: |
| Opc = NVPTX::SULD_1D_I64_ZERO; |
| break; |
| case NVPTXISD::Suld1DV2I8Zero: |
| Opc = NVPTX::SULD_1D_V2I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DV2I16Zero: |
| Opc = NVPTX::SULD_1D_V2I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DV2I32Zero: |
| Opc = NVPTX::SULD_1D_V2I32_ZERO; |
| break; |
| case NVPTXISD::Suld1DV2I64Zero: |
| Opc = NVPTX::SULD_1D_V2I64_ZERO; |
| break; |
| case NVPTXISD::Suld1DV4I8Zero: |
| Opc = NVPTX::SULD_1D_V4I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DV4I16Zero: |
| Opc = NVPTX::SULD_1D_V4I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DV4I32Zero: |
| Opc = NVPTX::SULD_1D_V4I32_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayI8Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayI16Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayI32Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_I32_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayI64Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_I64_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV2I8Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV2I16Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV2I32Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I32_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV2I64Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V2I64_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV4I8Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I8_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV4I16Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I16_ZERO; |
| break; |
| case NVPTXISD::Suld1DArrayV4I32Zero: |
| Opc = NVPTX::SULD_1D_ARRAY_V4I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DI8Zero: |
| Opc = NVPTX::SULD_2D_I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DI16Zero: |
| Opc = NVPTX::SULD_2D_I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DI32Zero: |
| Opc = NVPTX::SULD_2D_I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DI64Zero: |
| Opc = NVPTX::SULD_2D_I64_ZERO; |
| break; |
| case NVPTXISD::Suld2DV2I8Zero: |
| Opc = NVPTX::SULD_2D_V2I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DV2I16Zero: |
| Opc = NVPTX::SULD_2D_V2I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DV2I32Zero: |
| Opc = NVPTX::SULD_2D_V2I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DV2I64Zero: |
| Opc = NVPTX::SULD_2D_V2I64_ZERO; |
| break; |
| case NVPTXISD::Suld2DV4I8Zero: |
| Opc = NVPTX::SULD_2D_V4I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DV4I16Zero: |
| Opc = NVPTX::SULD_2D_V4I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DV4I32Zero: |
| Opc = NVPTX::SULD_2D_V4I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayI8Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayI16Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayI32Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayI64Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_I64_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV2I8Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV2I16Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV2I32Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I32_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV2I64Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V2I64_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV4I8Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I8_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV4I16Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I16_ZERO; |
| break; |
| case NVPTXISD::Suld2DArrayV4I32Zero: |
| Opc = NVPTX::SULD_2D_ARRAY_V4I32_ZERO; |
| break; |
| case NVPTXISD::Suld3DI8Zero: |
| Opc = NVPTX::SULD_3D_I8_ZERO; |
| break; |
| case NVPTXISD::Suld3DI16Zero: |
| Opc = NVPTX::SULD_3D_I16_ZERO; |
| break; |
| case NVPTXISD::Suld3DI32Zero: |
| Opc = NVPTX::SULD_3D_I32_ZERO; |
| break; |
| case NVPTXISD::Suld3DI64Zero: |
| Opc = NVPTX::SULD_3D_I64_ZERO; |
| break; |
| case NVPTXISD::Suld3DV2I8Zero: |
| Opc = NVPTX::SULD_3D_V2I8_ZERO; |
| break; |
| case NVPTXISD::Suld3DV2I16Zero: |
| Opc = NVPTX::SULD_3D_V2I16_ZERO; |
| break; |
| case NVPTXISD::Suld3DV2I32Zero: |
| Opc = NVPTX::SULD_3D_V2I32_ZERO; |
| break; |
| case NVPTXISD::Suld3DV2I64Zero: |
| Opc = NVPTX::SULD_3D_V2I64_ZERO; |
| break; |
| case NVPTXISD::Suld3DV4I8Zero: |
| Opc = NVPTX::SULD_3D_V4I8_ZERO; |
| break; |
| case NVPTXISD::Suld3DV4I16Zero: |
| Opc = NVPTX::SULD_3D_V4I16_ZERO; |
| break; |
| case NVPTXISD::Suld3DV4I32Zero: |
| Opc = NVPTX::SULD_3D_V4I32_ZERO; |
| break; |
| } |
| |
| // Copy over operands |
| SmallVector<SDValue, 8> Ops(N->op_begin() + 1, N->op_end()); |
| Ops.push_back(N->getOperand(0)); // Move chain to the back. |
| |
| ReplaceNode(N, CurDAG->getMachineNode(Opc, SDLoc(N), N->getVTList(), Ops)); |
| return true; |
| } |
| |
| |
| /// SelectBFE - Look for instruction sequences that can be made more efficient |
| /// by using the 'bfe' (bit-field extract) PTX instruction |
| bool NVPTXDAGToDAGISel::tryBFE(SDNode *N) { |
| SDLoc DL(N); |
| SDValue LHS = N->getOperand(0); |
| SDValue RHS = N->getOperand(1); |
| SDValue Len; |
| SDValue Start; |
| SDValue Val; |
| bool IsSigned = false; |
| |
| if (N->getOpcode() == ISD::AND) { |
| // Canonicalize the operands |
| // We want 'and %val, %mask' |
| if (isa<ConstantSDNode>(LHS) && !isa<ConstantSDNode>(RHS)) { |
| std::swap(LHS, RHS); |
| } |
| |
| ConstantSDNode *Mask = dyn_cast<ConstantSDNode>(RHS); |
| if (!Mask) { |
| // We need a constant mask on the RHS of the AND |
| return false; |
| } |
| |
| // Extract the mask bits |
| uint64_t MaskVal = Mask->getZExtValue(); |
| if (!isMask_64(MaskVal)) { |
| // We *could* handle shifted masks here, but doing so would require an |
| // 'and' operation to fix up the low-order bits so we would trade |
| // shr+and for bfe+and, which has the same throughput |
| return false; |
| } |
| |
| // How many bits are in our mask? |
| uint64_t NumBits = countTrailingOnes(MaskVal); |
| Len = CurDAG->getTargetConstant(NumBits, DL, MVT::i32); |
| |
| if (LHS.getOpcode() == ISD::SRL || LHS.getOpcode() == ISD::SRA) { |
| // We have a 'srl/and' pair, extract the effective start bit and length |
| Val = LHS.getNode()->getOperand(0); |
| Start = LHS.getNode()->getOperand(1); |
| ConstantSDNode *StartConst = dyn_cast<ConstantSDNode>(Start); |
| if (StartConst) { |
| uint64_t StartVal = StartConst->getZExtValue(); |
| // How many "good" bits do we have left? "good" is defined here as bits |
| // that exist in the original value, not shifted in. |
| uint64_t GoodBits = Start.getValueSizeInBits() - StartVal; |
| if (NumBits > GoodBits) { |
| // Do not handle the case where bits have been shifted in. In theory |
| // we could handle this, but the cost is likely higher than just |
| // emitting the srl/and pair. |
| return false; |
| } |
| Start = CurDAG->getTargetConstant(StartVal, DL, MVT::i32); |
| } else { |
| // Do not handle the case where the shift amount (can be zero if no srl |
| // was found) is not constant. We could handle this case, but it would |
| // require run-time logic that would be more expensive than just |
| // emitting the srl/and pair. |
| return false; |
| } |
| } else { |
| // Do not handle the case where the LHS of the and is not a shift. While |
| // it would be trivial to handle this case, it would just transform |
| // 'and' -> 'bfe', but 'and' has higher-throughput. |
| return false; |
| } |
| } else if (N->getOpcode() == ISD::SRL || N->getOpcode() == ISD::SRA) { |
| if (LHS->getOpcode() == ISD::AND) { |
| ConstantSDNode *ShiftCnst = dyn_cast<ConstantSDNode>(RHS); |
| if (!ShiftCnst) { |
| // Shift amount must be constant |
| return false; |
| } |
| |
| uint64_t ShiftAmt = ShiftCnst->getZExtValue(); |
| |
| SDValue AndLHS = LHS->getOperand(0); |
| SDValue AndRHS = LHS->getOperand(1); |
| |
| // Canonicalize the AND to have the mask on the RHS |
| if (isa<ConstantSDNode>(AndLHS)) { |
| std::swap(AndLHS, AndRHS); |
| } |
| |
| ConstantSDNode *MaskCnst = dyn_cast<ConstantSDNode>(AndRHS); |
| if (!MaskCnst) { |
| // Mask must be constant |
| return false; |
| } |
| |
| uint64_t MaskVal = MaskCnst->getZExtValue(); |
| uint64_t NumZeros; |
| uint64_t NumBits; |
| if (isMask_64(MaskVal)) { |
| NumZeros = 0; |
| // The number of bits in the result bitfield will be the number of |
| // trailing ones (the AND) minus the number of bits we shift off |
| NumBits = countTrailingOnes(MaskVal) - ShiftAmt; |
| } else if (isShiftedMask_64(MaskVal)) { |
| NumZeros = countTrailingZeros(MaskVal); |
| unsigned NumOnes = countTrailingOnes(MaskVal >> NumZeros); |
| // The number of bits in the result bitfield will be the number of |
| // trailing zeros plus the number of set bits in the mask minus the |
| // number of bits we shift off |
| NumBits = NumZeros + NumOnes - ShiftAmt; |
| } else { |
| // This is not a mask we can handle |
| return false; |
| } |
| |
| if (ShiftAmt < NumZeros) { |
| // Handling this case would require extra logic that would make this |
| // transformation non-profitable |
| return false; |
| } |
| |
| Val = AndLHS; |
| Start = CurDAG->getTargetConstant(ShiftAmt, DL, MVT::i32); |
| Len = CurDAG->getTargetConstant(NumBits, DL, MVT::i32); |
| } else if (LHS->getOpcode() == ISD::SHL) { |
| // Here, we have a pattern like: |
| // |
| // (sra (shl val, NN), MM) |
| // or |
| // (srl (shl val, NN), MM) |
| // |
| // If MM >= NN, we can efficiently optimize this with bfe |
| Val = LHS->getOperand(0); |
| |
| SDValue ShlRHS = LHS->getOperand(1); |
| ConstantSDNode *ShlCnst = dyn_cast<ConstantSDNode>(ShlRHS); |
| if (!ShlCnst) { |
| // Shift amount must be constant |
| return false; |
| } |
| uint64_t InnerShiftAmt = ShlCnst->getZExtValue(); |
| |
| SDValue ShrRHS = RHS; |
| ConstantSDNode *ShrCnst = dyn_cast<ConstantSDNode>(ShrRHS); |
| if (!ShrCnst) { |
| // Shift amount must be constant |
| return false; |
| } |
| uint64_t OuterShiftAmt = ShrCnst->getZExtValue(); |
| |
| // To avoid extra codegen and be profitable, we need Outer >= Inner |
| if (OuterShiftAmt < InnerShiftAmt) { |
| return false; |
| } |
| |
| // If the outer shift is more than the type size, we have no bitfield to |
| // extract (since we also check that the inner shift is <= the outer shift |
| // then this also implies that the inner shift is < the type size) |
| if (OuterShiftAmt >= Val.getValueSizeInBits()) { |
| return false; |
| } |
| |
| Start = CurDAG->getTargetConstant(OuterShiftAmt - InnerShiftAmt, DL, |
| MVT::i32); |
| Len = CurDAG->getTargetConstant(Val.getValueSizeInBits() - OuterShiftAmt, |
| DL, MVT::i32); |
| |
| if (N->getOpcode() == ISD::SRA) { |
| // If we have a arithmetic right shift, we need to use the signed bfe |
| // variant |
| IsSigned = true; |
| } |
| } else { |
| // No can do... |
| return false; |
| } |
| } else { |
| // No can do... |
| return false; |
| } |
| |
| |
| unsigned Opc; |
| // For the BFE operations we form here from "and" and "srl", always use the |
| // unsigned variants. |
| if (Val.getValueType() == MVT::i32) { |
| if (IsSigned) { |
| Opc = NVPTX::BFE_S32rii; |
| } else { |
| Opc = NVPTX::BFE_U32rii; |
| } |
| } else if (Val.getValueType() == MVT::i64) { |
| if (IsSigned) { |
| Opc = NVPTX::BFE_S64rii; |
| } else { |
| Opc = NVPTX::BFE_U64rii; |
| } |
| } else { |
| // We cannot handle this type |
| return false; |
| } |
| |
| SDValue Ops[] = { |
| Val, Start, Len |
| }; |
| |
| ReplaceNode(N, CurDAG->getMachineNode(Opc, DL, N->getVTList(), Ops)); |
| return true; |
| } |
| |
| // SelectDirectAddr - Match a direct address for DAG. |
| // A direct address could be a globaladdress or externalsymbol. |
| bool NVPTXDAGToDAGISel::SelectDirectAddr(SDValue N, SDValue &Address) { |
| // Return true if TGA or ES. |
| if (N.getOpcode() == ISD::TargetGlobalAddress || |
| N.getOpcode() == ISD::TargetExternalSymbol) { |
| Address = N; |
| return true; |
| } |
| if (N.getOpcode() == NVPTXISD::Wrapper) { |
| Address = N.getOperand(0); |
| return true; |
| } |
| // addrspacecast(MoveParam(arg_symbol) to addrspace(PARAM)) -> arg_symbol |
| if (AddrSpaceCastSDNode *CastN = dyn_cast<AddrSpaceCastSDNode>(N)) { |
| if (CastN->getSrcAddressSpace() == ADDRESS_SPACE_GENERIC && |
| CastN->getDestAddressSpace() == ADDRESS_SPACE_PARAM && |
| CastN->getOperand(0).getOpcode() == NVPTXISD::MoveParam) |
| return SelectDirectAddr(CastN->getOperand(0).getOperand(0), Address); |
| } |
| return false; |
| } |
| |
| // symbol+offset |
| bool NVPTXDAGToDAGISel::SelectADDRsi_imp( |
| SDNode *OpNode, SDValue Addr, SDValue &Base, SDValue &Offset, MVT mvt) { |
| if (Addr.getOpcode() == ISD::ADD) { |
| if (ConstantSDNode *CN = dyn_cast<ConstantSDNode>(Addr.getOperand(1))) { |
| SDValue base = Addr.getOperand(0); |
| if (SelectDirectAddr(base, Base)) { |
| Offset = CurDAG->getTargetConstant(CN->getZExtValue(), SDLoc(OpNode), |
| mvt); |
| return true; |
| } |
| } |
| } |
| return false; |
| } |
| |
| // symbol+offset |
| bool NVPTXDAGToDAGISel::SelectADDRsi(SDNode *OpNode, SDValue Addr, |
| SDValue &Base, SDValue &Offset) { |
| return SelectADDRsi_imp(OpNode, Addr, Base, Offset, MVT::i32); |
| } |
| |
| // symbol+offset |
| bool NVPTXDAGToDAGISel::SelectADDRsi64(SDNode *OpNode, SDValue Addr, |
| SDValue &Base, SDValue &Offset) { |
| return SelectADDRsi_imp(OpNode, Addr, Base, Offset, MVT::i64); |
| } |
| |
| // register+offset |
| bool NVPTXDAGToDAGISel::SelectADDRri_imp( |
| SDNode *OpNode, SDValue Addr, SDValue &Base, SDValue &Offset, MVT mvt) { |
| if (FrameIndexSDNode *FIN = dyn_cast<FrameIndexSDNode>(Addr)) { |
| Base = CurDAG->getTargetFrameIndex(FIN->getIndex(), mvt); |
| Offset = CurDAG->getTargetConstant(0, SDLoc(OpNode), mvt); |
| return true; |
| } |
| if (Addr.getOpcode() == ISD::TargetExternalSymbol || |
| Addr.getOpcode() == ISD::TargetGlobalAddress) |
| return false; // direct calls. |
| |
| if (Addr.getOpcode() == ISD::ADD) { |
| if (SelectDirectAddr(Addr.getOperand(0), Addr)) { |
| return false; |
| } |
| if (ConstantSDNode *CN = dyn_cast<ConstantSDNode>(Addr.getOperand(1))) { |
| if (FrameIndexSDNode *FIN = |
| dyn_cast<FrameIndexSDNode>(Addr.getOperand(0))) |
| // Constant offset from frame ref. |
| Base = CurDAG->getTargetFrameIndex(FIN->getIndex(), mvt); |
| else |
| Base = Addr.getOperand(0); |
| Offset = CurDAG->getTargetConstant(CN->getZExtValue(), SDLoc(OpNode), |
| mvt); |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| // register+offset |
| bool NVPTXDAGToDAGISel::SelectADDRri(SDNode *OpNode, SDValue Addr, |
| SDValue &Base, SDValue &Offset) { |
| return SelectADDRri_imp(OpNode, Addr, Base, Offset, MVT::i32); |
| } |
| |
| // register+offset |
| bool NVPTXDAGToDAGISel::SelectADDRri64(SDNode *OpNode, SDValue Addr, |
| SDValue &Base, SDValue &Offset) { |
| return SelectADDRri_imp(OpNode, Addr, Base, Offset, MVT::i64); |
| } |
| |
| bool NVPTXDAGToDAGISel::ChkMemSDNodeAddressSpace(SDNode *N, |
| unsigned int spN) const { |
| const Value *Src = nullptr; |
| if (MemSDNode *mN = dyn_cast<MemSDNode>(N)) { |
| if (spN == 0 && mN->getMemOperand()->getPseudoValue()) |
| return true; |
| Src = mN->getMemOperand()->getValue(); |
| } |
| if (!Src) |
| return false; |
| if (auto *PT = dyn_cast<PointerType>(Src->getType())) |
| return (PT->getAddressSpace() == spN); |
| return false; |
| } |
| |
| /// SelectInlineAsmMemoryOperand - Implement addressing mode selection for |
| /// inline asm expressions. |
| bool NVPTXDAGToDAGISel::SelectInlineAsmMemoryOperand( |
| const SDValue &Op, unsigned ConstraintID, std::vector<SDValue> &OutOps) { |
| SDValue Op0, Op1; |
| switch (ConstraintID) { |
| default: |
| return true; |
| case InlineAsm::Constraint_m: // memory |
| if (SelectDirectAddr(Op, Op0)) { |
| OutOps.push_back(Op0); |
| OutOps.push_back(CurDAG->getTargetConstant(0, SDLoc(Op), MVT::i32)); |
| return false; |
| } |
| if (SelectADDRri(Op.getNode(), Op, Op0, Op1)) { |
| OutOps.push_back(Op0); |
| OutOps.push_back(Op1); |
| return false; |
| } |
| break; |
| } |
| return true; |
| } |
| |
| /// GetConvertOpcode - Returns the CVT_ instruction opcode that implements a |
| /// conversion from \p SrcTy to \p DestTy. |
| unsigned NVPTXDAGToDAGISel::GetConvertOpcode(MVT DestTy, MVT SrcTy, |
| bool IsSigned) { |
| switch (SrcTy.SimpleTy) { |
| default: |
| llvm_unreachable("Unhandled source type"); |
| case MVT::i8: |
| switch (DestTy.SimpleTy) { |
| default: |
| llvm_unreachable("Unhandled dest type"); |
| case MVT::i16: |
| return IsSigned ? NVPTX::CVT_s16_s8 : NVPTX::CVT_u16_u8; |
| case MVT::i32: |
| return IsSigned ? NVPTX::CVT_s32_s8 : NVPTX::CVT_u32_u8; |
| case MVT::i64: |
| return IsSigned ? NVPTX::CVT_s64_s8 : NVPTX::CVT_u64_u8; |
| } |
| case MVT::i16: |
| switch (DestTy.SimpleTy) { |
| default: |
| llvm_unreachable("Unhandled dest type"); |
| case MVT::i8: |
| return IsSigned ? NVPTX::CVT_s8_s16 : NVPTX::CVT_u8_u16; |
| case MVT::i32: |
| return IsSigned ? NVPTX::CVT_s32_s16 : NVPTX::CVT_u32_u16; |
| case MVT::i64: |
| return IsSigned ? NVPTX::CVT_s64_s16 : NVPTX::CVT_u64_u16; |
| } |
| case MVT::i32: |
| switch (DestTy.SimpleTy) { |
| default: |
| llvm_unreachable("Unhandled dest type"); |
| case MVT::i8: |
| return IsSigned ? NVPTX::CVT_s8_s32 : NVPTX::CVT_u8_u32; |
| case MVT::i16: |
| return IsSigned ? NVPTX::CVT_s16_s32 : NVPTX::CVT_u16_u32; |
| case MVT::i64: |
| return IsSigned ? NVPTX::CVT_s64_s32 : NVPTX::CVT_u64_u32; |
| } |
| case MVT::i64: |
| switch (DestTy.SimpleTy) { |
| default: |
| llvm_unreachable("Unhandled dest type"); |
| case MVT::i8: |
| return IsSigned ? NVPTX::CVT_s8_s64 : NVPTX::CVT_u8_u64; |
| case MVT::i16: |
| return IsSigned ? NVPTX::CVT_s16_s64 : NVPTX::CVT_u16_u64; |
| case MVT::i32: |
| return IsSigned ? NVPTX::CVT_s32_s64 : NVPTX::CVT_u32_u64; |
| } |
| } |
| } |