| //===- subzero/src/IceTargetLoweringX8664Traits.h - x86-64 traits -*- C++ -*-=// |
| // |
| // The Subzero Code Generator |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| /// |
| /// \file |
| /// \brief Declares the X8664 Target Lowering Traits. |
| /// |
| //===----------------------------------------------------------------------===// |
| |
| #ifndef SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H |
| #define SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H |
| |
| #include "IceAssembler.h" |
| #include "IceConditionCodesX8664.h" |
| #include "IceDefs.h" |
| #include "IceInst.h" |
| #include "IceInstX8664.def" |
| #include "IceOperand.h" |
| #include "IceRegistersX8664.h" |
| #include "IceTargetLowering.h" |
| #include "IceTargetLoweringX8664.def" |
| #include "IceTargetLoweringX86RegClass.h" |
| |
| #include <array> |
| #include <initializer_list> |
| |
| namespace Ice { |
| |
| namespace X8664 { |
| using namespace ::Ice::X86; |
| |
| template <class TraitsType> class AssemblerX86Base; |
| template <class TraitsType> struct Insts; |
| template <class TraitsType> class TargetX86Base; |
| |
| class TargetX8664; |
| |
| struct TargetX8664Traits { |
| //---------------------------------------------------------------------------- |
| // ______ ______ __ __ |
| // /\ __ \/\ ___\/\ "-./ \ |
| // \ \ __ \ \___ \ \ \-./\ \ |
| // \ \_\ \_\/\_____\ \_\ \ \_\ |
| // \/_/\/_/\/_____/\/_/ \/_/ |
| // |
| //---------------------------------------------------------------------------- |
| static constexpr ::Ice::Assembler::AssemblerKind AsmKind = |
| ::Ice::Assembler::Asm_X8664; |
| |
| static constexpr bool Is64Bit = true; |
| static constexpr bool HasPopa = false; |
| static constexpr bool HasPusha = false; |
| static constexpr bool UsesX87 = false; |
| static constexpr ::Ice::RegX8664::GPRRegister Last8BitGPR = |
| ::Ice::RegX8664::GPRRegister::Encoded_Reg_r15d; |
| |
| enum ScaleFactor { TIMES_1 = 0, TIMES_2 = 1, TIMES_4 = 2, TIMES_8 = 3 }; |
| |
| using GPRRegister = ::Ice::RegX8664::GPRRegister; |
| using ByteRegister = ::Ice::RegX8664::ByteRegister; |
| using XmmRegister = ::Ice::RegX8664::XmmRegister; |
| |
| using Cond = ::Ice::CondX8664; |
| |
| using RegisterSet = ::Ice::RegX8664; |
| static constexpr RegisterSet::AllRegisters StackPtr = RegX8664::Reg_rsp; |
| static constexpr RegisterSet::AllRegisters FramePtr = RegX8664::Reg_rbp; |
| static constexpr GPRRegister Encoded_Reg_Accumulator = |
| RegX8664::Encoded_Reg_eax; |
| static constexpr GPRRegister Encoded_Reg_Counter = RegX8664::Encoded_Reg_ecx; |
| static constexpr FixupKind FK_PcRel = llvm::ELF::R_X86_64_PC32; |
| static constexpr FixupKind FK_Abs = llvm::ELF::R_X86_64_32S; |
| static constexpr FixupKind FK_Gotoff = llvm::ELF::R_X86_64_GOTOFF64; |
| static constexpr FixupKind FK_GotPC = llvm::ELF::R_X86_64_GOTPC32; |
| |
| class Operand { |
| public: |
| enum RexBits { |
| RexNone = 0x00, |
| RexBase = 0x40, |
| RexW = RexBase | (1 << 3), |
| RexR = RexBase | (1 << 2), |
| RexX = RexBase | (1 << 1), |
| RexB = RexBase | (1 << 0), |
| }; |
| |
| protected: |
| // Needed by subclass Address. |
| Operand() = default; |
| |
| public: |
| Operand(const Operand &) = default; |
| Operand(Operand &&) = default; |
| Operand &operator=(const Operand &) = default; |
| Operand &operator=(Operand &&) = default; |
| |
| uint8_t mod() const { return (encoding_at(0) >> 6) & 3; } |
| |
| uint8_t rexX() const { return (rex_ & RexX) != RexX ? RexNone : RexX; } |
| uint8_t rexB() const { return (rex_ & RexB) != RexB ? RexNone : RexB; } |
| |
| GPRRegister rm() const { |
| return static_cast<GPRRegister>((rexB() != 0 ? 0x08 : 0) | |
| (encoding_at(0) & 7)); |
| } |
| |
| ScaleFactor scale() const { |
| return static_cast<ScaleFactor>((encoding_at(1) >> 6) & 3); |
| } |
| |
| GPRRegister index() const { |
| return static_cast<GPRRegister>((rexX() != 0 ? 0x08 : 0) | |
| ((encoding_at(1) >> 3) & 7)); |
| } |
| |
| GPRRegister base() const { |
| return static_cast<GPRRegister>((rexB() != 0 ? 0x08 : 0) | |
| (encoding_at(1) & 7)); |
| } |
| |
| int8_t disp8() const { |
| assert(length_ >= 2); |
| return static_cast<int8_t>(encoding_[length_ - 1]); |
| } |
| |
| AssemblerFixup *fixup() const { return fixup_; } |
| |
| protected: |
| void SetModRM(int mod, GPRRegister rm) { |
| assert((mod & ~3) == 0); |
| encoding_[0] = (mod << 6) | (rm & 0x07); |
| rex_ = (rm & 0x08) ? RexB : RexNone; |
| length_ = 1; |
| } |
| |
| void SetSIB(ScaleFactor scale, GPRRegister index, GPRRegister base) { |
| assert(length_ == 1); |
| assert((scale & ~3) == 0); |
| encoding_[1] = (scale << 6) | ((index & 0x07) << 3) | (base & 0x07); |
| rex_ = |
| ((base & 0x08) ? RexB : RexNone) | ((index & 0x08) ? RexX : RexNone); |
| length_ = 2; |
| } |
| |
| void SetDisp8(int8_t disp) { |
| assert(length_ == 1 || length_ == 2); |
| encoding_[length_++] = static_cast<uint8_t>(disp); |
| } |
| |
| void SetDisp32(int32_t disp) { |
| assert(length_ == 1 || length_ == 2); |
| intptr_t disp_size = sizeof(disp); |
| memmove(&encoding_[length_], &disp, disp_size); |
| length_ += disp_size; |
| } |
| |
| void SetFixup(AssemblerFixup *fixup) { fixup_ = fixup; } |
| |
| private: |
| AssemblerFixup *fixup_ = nullptr; |
| uint8_t rex_ = 0; |
| uint8_t encoding_[6]; |
| uint8_t length_ = 0; |
| |
| explicit Operand(GPRRegister reg) : fixup_(nullptr) { SetModRM(3, reg); } |
| |
| /// Get the operand encoding byte at the given index. |
| uint8_t encoding_at(intptr_t index) const { |
| assert(index >= 0 && index < length_); |
| return encoding_[index]; |
| } |
| |
| /// Returns whether or not this operand is really the given register in |
| /// disguise. Used from the assembler to generate better encodings. |
| bool IsRegister(GPRRegister reg) const { |
| return ((encoding_[0] & 0xF8) == |
| 0xC0) // Addressing mode is register only. |
| && |
| (rm() == reg); // Register codes match. |
| } |
| |
| friend class AssemblerX86Base<TargetX8664Traits>; |
| }; |
| |
| class Address : public Operand { |
| Address() = default; |
| |
| public: |
| Address(const Address &) = default; |
| Address(Address &&) = default; |
| Address &operator=(const Address &) = default; |
| Address &operator=(Address &&) = default; |
| |
| Address(GPRRegister Base, int32_t Disp, AssemblerFixup *Fixup) { |
| if (Fixup == nullptr && Disp == 0 && |
| (Base & 7) != RegX8664::Encoded_Reg_rbp) { |
| SetModRM(0, Base); |
| if ((Base & 7) == RegX8664::Encoded_Reg_rsp) |
| SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); |
| } else if (Fixup == nullptr && Utils::IsInt(8, Disp)) { |
| SetModRM(1, Base); |
| if ((Base & 7) == RegX8664::Encoded_Reg_rsp) |
| SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); |
| SetDisp8(Disp); |
| } else { |
| SetModRM(2, Base); |
| if ((Base & 7) == RegX8664::Encoded_Reg_rsp) |
| SetSIB(TIMES_1, RegX8664::Encoded_Reg_rsp, Base); |
| SetDisp32(Disp); |
| if (Fixup) |
| SetFixup(Fixup); |
| } |
| } |
| |
| Address(GPRRegister Index, ScaleFactor Scale, int32_t Disp, |
| AssemblerFixup *Fixup) { |
| assert(Index != RegX8664::Encoded_Reg_rsp); // Illegal addressing mode. |
| SetModRM(0, RegX8664::Encoded_Reg_rsp); |
| SetSIB(Scale, Index, RegX8664::Encoded_Reg_rbp); |
| SetDisp32(Disp); |
| if (Fixup) |
| SetFixup(Fixup); |
| } |
| |
| Address(GPRRegister Base, GPRRegister Index, ScaleFactor Scale, |
| int32_t Disp, AssemblerFixup *Fixup) { |
| assert(Index != RegX8664::Encoded_Reg_rsp); // Illegal addressing mode. |
| if (Fixup == nullptr && Disp == 0 && |
| (Base & 7) != RegX8664::Encoded_Reg_rbp) { |
| SetModRM(0, RegX8664::Encoded_Reg_rsp); |
| SetSIB(Scale, Index, Base); |
| } else if (Fixup == nullptr && Utils::IsInt(8, Disp)) { |
| SetModRM(1, RegX8664::Encoded_Reg_rsp); |
| SetSIB(Scale, Index, Base); |
| SetDisp8(Disp); |
| } else { |
| SetModRM(2, RegX8664::Encoded_Reg_rsp); |
| SetSIB(Scale, Index, Base); |
| SetDisp32(Disp); |
| if (Fixup) |
| SetFixup(Fixup); |
| } |
| } |
| |
| /// Generate a RIP-relative address expression on x86-64. |
| static Address RipRelative(RelocOffsetT Offset, AssemblerFixup *Fixup) { |
| assert(Fixup != nullptr); |
| assert(Fixup->kind() == FK_PcRel); |
| Address NewAddress; |
| NewAddress.SetModRM(0x0, RegX8664::Encoded_Reg_rbp); |
| |
| // Use the Offset in the displacement for now. If we decide to process |
| // fixups later, we'll need to patch up the emitted displacement. |
| NewAddress.SetDisp32(Offset); |
| if (Fixup) |
| NewAddress.SetFixup(Fixup); |
| |
| return NewAddress; |
| } |
| |
| /// Generate an absolute address. |
| static Address Absolute(RelocOffsetT Addr) { |
| Address NewAddress; |
| NewAddress.SetModRM(0x0, RegX8664::Encoded_Reg_rsp); |
| static constexpr ScaleFactor NoScale = TIMES_1; |
| NewAddress.SetSIB(NoScale, RegX8664::Encoded_Reg_rsp, |
| RegX8664::Encoded_Reg_rbp); |
| NewAddress.SetDisp32(Addr); |
| return NewAddress; |
| } |
| |
| static Address ofConstPool(Assembler *Asm, const Constant *Imm) { |
| // TODO(jpp): ??? |
| AssemblerFixup *Fixup = Asm->createFixup(FK_Abs, Imm); |
| const RelocOffsetT Offset = 4; |
| return Address::RipRelative(Offset, Fixup); |
| } |
| }; |
| |
| //---------------------------------------------------------------------------- |
| // __ ______ __ __ ______ ______ __ __ __ ______ |
| // /\ \ /\ __ \/\ \ _ \ \/\ ___\/\ == \/\ \/\ "-.\ \/\ ___\ |
| // \ \ \___\ \ \/\ \ \ \/ ".\ \ \ __\\ \ __<\ \ \ \ \-. \ \ \__ \ |
| // \ \_____\ \_____\ \__/".~\_\ \_____\ \_\ \_\ \_\ \_\\"\_\ \_____\ |
| // \/_____/\/_____/\/_/ \/_/\/_____/\/_/ /_/\/_/\/_/ \/_/\/_____/ |
| // |
| //---------------------------------------------------------------------------- |
| enum InstructionSet { |
| Begin, |
| // SSE2 is the PNaCl baseline instruction set. |
| SSE2 = Begin, |
| SSE4_1, |
| End |
| }; |
| |
| static const char *TargetName; |
| static constexpr Type WordType = IceType_i64; |
| |
| static const char *getRegName(RegNumT RegNum) { |
| static const char *const RegNames[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| name, |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| return RegNames[RegNum]; |
| } |
| |
| static GPRRegister getEncodedGPR(RegNumT RegNum) { |
| static const GPRRegister GPRRegs[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| GPRRegister(isGPR ? encode : GPRRegister::Encoded_Not_GPR), |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| assert(GPRRegs[RegNum] != GPRRegister::Encoded_Not_GPR); |
| return GPRRegs[RegNum]; |
| } |
| |
| static ByteRegister getEncodedByteReg(RegNumT RegNum) { |
| static const ByteRegister ByteRegs[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| ByteRegister(is8 ? encode : ByteRegister::Encoded_Not_ByteReg), |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| assert(ByteRegs[RegNum] != ByteRegister::Encoded_Not_ByteReg); |
| return ByteRegs[RegNum]; |
| } |
| |
| static XmmRegister getEncodedXmm(RegNumT RegNum) { |
| static const XmmRegister XmmRegs[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| XmmRegister(isXmm ? encode : XmmRegister::Encoded_Not_Xmm), |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| assert(XmmRegs[RegNum] != XmmRegister::Encoded_Not_Xmm); |
| return XmmRegs[RegNum]; |
| } |
| |
| static uint32_t getEncoding(RegNumT RegNum) { |
| static const uint32_t Encoding[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| encode, |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| return Encoding[RegNum]; |
| } |
| |
| static inline RegNumT getBaseReg(RegNumT RegNum) { |
| static const RegNumT BaseRegs[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| RegisterSet::base, |
| REGX8664_TABLE |
| #undef X |
| }; |
| RegNum.assertIsValid(); |
| return BaseRegs[RegNum]; |
| } |
| |
| private: |
| static RegNumT getFirstGprForType(Type Ty) { |
| switch (Ty) { |
| default: |
| llvm_unreachable("Invalid type for GPR."); |
| case IceType_i1: |
| case IceType_i8: |
| return RegisterSet::Reg_al; |
| case IceType_i16: |
| return RegisterSet::Reg_ax; |
| case IceType_i32: |
| return RegisterSet::Reg_eax; |
| case IceType_i64: |
| return RegisterSet::Reg_rax; |
| } |
| } |
| |
| public: |
| static RegNumT getGprForType(Type Ty, RegNumT RegNum) { |
| assert(RegNum.hasValue()); |
| |
| if (!isScalarIntegerType(Ty)) { |
| return RegNum; |
| } |
| |
| assert(Ty == IceType_i1 || Ty == IceType_i8 || Ty == IceType_i16 || |
| Ty == IceType_i32 || Ty == IceType_i64); |
| |
| if (RegNum == RegisterSet::Reg_ah) { |
| assert(Ty == IceType_i8); |
| return RegNum; |
| } |
| |
| assert(RegNum != RegisterSet::Reg_bh); |
| assert(RegNum != RegisterSet::Reg_ch); |
| assert(RegNum != RegisterSet::Reg_dh); |
| |
| const RegNumT FirstGprForType = getFirstGprForType(Ty); |
| |
| switch (RegNum) { |
| default: |
| llvm::report_fatal_error("Unknown register."); |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| case RegisterSet::val: { \ |
| if (!isGPR) \ |
| return RegisterSet::val; \ |
| assert((is64) || (is32) || (is16) || (is8) || \ |
| getBaseReg(RegisterSet::val) == RegisterSet::Reg_rsp); \ |
| constexpr RegisterSet::AllRegisters FirstGprWithRegNumSize = \ |
| ((is64) || RegisterSet::val == RegisterSet::Reg_rsp) \ |
| ? RegisterSet::Reg_rax \ |
| : (((is32) || RegisterSet::val == RegisterSet::Reg_esp) \ |
| ? RegisterSet::Reg_eax \ |
| : (((is16) || RegisterSet::val == RegisterSet::Reg_sp) \ |
| ? RegisterSet::Reg_ax \ |
| : RegisterSet::Reg_al)); \ |
| const auto NewRegNum = \ |
| RegNumT::fixme(RegNum - FirstGprWithRegNumSize + FirstGprForType); \ |
| assert(getBaseReg(RegNum) == getBaseReg(NewRegNum) && \ |
| "Error involving " #val); \ |
| return NewRegNum; \ |
| } |
| REGX8664_TABLE |
| #undef X |
| } |
| } |
| |
| private: |
| /// SizeOf is used to obtain the size of an initializer list as a constexpr |
| /// expression. This is only needed until our C++ library is updated to |
| /// C++ 14 -- which defines constexpr members to std::initializer_list. |
| class SizeOf { |
| SizeOf(const SizeOf &) = delete; |
| SizeOf &operator=(const SizeOf &) = delete; |
| |
| public: |
| constexpr SizeOf() : Size(0) {} |
| template <typename... T> |
| explicit constexpr SizeOf(T...) |
| : Size(length<T...>::value) {} |
| constexpr SizeT size() const { return Size; } |
| |
| private: |
| template <typename T, typename... U> struct length { |
| static constexpr std::size_t value = 1 + length<U...>::value; |
| }; |
| |
| template <typename T> struct length<T> { |
| static constexpr std::size_t value = 1; |
| }; |
| |
| const std::size_t Size; |
| }; |
| |
| public: |
| static void initRegisterSet( |
| const ::Ice::ClFlags &Flags, |
| std::array<SmallBitVector, RCX86_NUM> *TypeToRegisterSet, |
| std::array<SmallBitVector, RegisterSet::Reg_NUM> *RegisterAliases) { |
| SmallBitVector IntegerRegistersI64(RegisterSet::Reg_NUM); |
| SmallBitVector IntegerRegistersI32(RegisterSet::Reg_NUM); |
| SmallBitVector IntegerRegistersI16(RegisterSet::Reg_NUM); |
| SmallBitVector IntegerRegistersI8(RegisterSet::Reg_NUM); |
| SmallBitVector FloatRegisters(RegisterSet::Reg_NUM); |
| SmallBitVector VectorRegisters(RegisterSet::Reg_NUM); |
| SmallBitVector Trunc64To8Registers(RegisterSet::Reg_NUM); |
| SmallBitVector Trunc32To8Registers(RegisterSet::Reg_NUM); |
| SmallBitVector Trunc16To8Registers(RegisterSet::Reg_NUM); |
| SmallBitVector Trunc8RcvrRegisters(RegisterSet::Reg_NUM); |
| SmallBitVector AhRcvrRegisters(RegisterSet::Reg_NUM); |
| SmallBitVector InvalidRegisters(RegisterSet::Reg_NUM); |
| |
| static constexpr struct { |
| uint16_t Val; |
| unsigned IsReservedWhenSandboxing : 1; |
| unsigned Is64 : 1; |
| unsigned Is32 : 1; |
| unsigned Is16 : 1; |
| unsigned Is8 : 1; |
| unsigned IsXmm : 1; |
| unsigned Is64To8 : 1; |
| unsigned Is32To8 : 1; |
| unsigned Is16To8 : 1; |
| unsigned IsTrunc8Rcvr : 1; |
| unsigned IsAhRcvr : 1; |
| #define NUM_ALIASES_BITS 2 |
| SizeT NumAliases : (NUM_ALIASES_BITS + 1); |
| uint16_t Aliases[1 << NUM_ALIASES_BITS]; |
| #undef NUM_ALIASES_BITS |
| } X8664RegTable[RegisterSet::Reg_NUM] = { |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| { \ |
| RegisterSet::val, sboxres, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, (SizeOf aliases).size(), aliases, \ |
| } \ |
| , |
| REGX8664_TABLE |
| #undef X |
| }; |
| |
| const bool NeedSandboxing = Flags.getUseSandboxing(); |
| for (SizeT ii = 0; ii < llvm::array_lengthof(X8664RegTable); ++ii) { |
| const auto &Entry = X8664RegTable[ii]; |
| // Even though the register is disabled for register allocation, it might |
| // still be used by the Target Lowering (e.g., base pointer), so the |
| // register alias table still needs to be defined. |
| (*RegisterAliases)[Entry.Val].resize(RegisterSet::Reg_NUM); |
| for (int J = 0; J < Entry.NumAliases; ++J) { |
| SizeT Alias = Entry.Aliases[J]; |
| assert(!(*RegisterAliases)[Entry.Val][Alias] && "Duplicate alias"); |
| (*RegisterAliases)[Entry.Val].set(Alias); |
| } |
| |
| (*RegisterAliases)[Entry.Val].set(Entry.Val); |
| const bool DisabledRegister = |
| NeedSandboxing && Entry.IsReservedWhenSandboxing; |
| if (DisabledRegister) { |
| continue; |
| } |
| (IntegerRegistersI64)[Entry.Val] = Entry.Is64; |
| (IntegerRegistersI32)[Entry.Val] = Entry.Is32; |
| (IntegerRegistersI16)[Entry.Val] = Entry.Is16; |
| (IntegerRegistersI8)[Entry.Val] = Entry.Is8; |
| (FloatRegisters)[Entry.Val] = Entry.IsXmm; |
| (VectorRegisters)[Entry.Val] = Entry.IsXmm; |
| (Trunc64To8Registers)[Entry.Val] = Entry.Is64To8; |
| (Trunc32To8Registers)[Entry.Val] = Entry.Is32To8; |
| (Trunc16To8Registers)[Entry.Val] = Entry.Is16To8; |
| (Trunc8RcvrRegisters)[Entry.Val] = Entry.IsTrunc8Rcvr; |
| (AhRcvrRegisters)[Entry.Val] = Entry.IsAhRcvr; |
| } |
| |
| (*TypeToRegisterSet)[RC_void] = InvalidRegisters; |
| (*TypeToRegisterSet)[RC_i1] = IntegerRegistersI8; |
| (*TypeToRegisterSet)[RC_i8] = IntegerRegistersI8; |
| (*TypeToRegisterSet)[RC_i16] = IntegerRegistersI16; |
| (*TypeToRegisterSet)[RC_i32] = IntegerRegistersI32; |
| (*TypeToRegisterSet)[RC_i64] = IntegerRegistersI64; |
| (*TypeToRegisterSet)[RC_f32] = FloatRegisters; |
| (*TypeToRegisterSet)[RC_f64] = FloatRegisters; |
| (*TypeToRegisterSet)[RC_v4i1] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v8i1] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v16i1] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v16i8] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v8i16] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v4i32] = VectorRegisters; |
| (*TypeToRegisterSet)[RC_v4f32] = VectorRegisters; |
| (*TypeToRegisterSet)[RCX86_Is64To8] = Trunc64To8Registers; |
| (*TypeToRegisterSet)[RCX86_Is32To8] = Trunc32To8Registers; |
| (*TypeToRegisterSet)[RCX86_Is16To8] = Trunc16To8Registers; |
| (*TypeToRegisterSet)[RCX86_IsTrunc8Rcvr] = Trunc8RcvrRegisters; |
| (*TypeToRegisterSet)[RCX86_IsAhRcvr] = AhRcvrRegisters; |
| } |
| |
| static SmallBitVector getRegisterSet(const ::Ice::ClFlags &Flags, |
| TargetLowering::RegSetMask Include, |
| TargetLowering::RegSetMask Exclude) { |
| SmallBitVector Registers(RegisterSet::Reg_NUM); |
| |
| const bool NeedSandboxing = Flags.getUseSandboxing(); |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| if (!NeedSandboxing || !(sboxres)) { \ |
| if (scratch && (Include & ::Ice::TargetLowering::RegSet_CallerSave)) \ |
| Registers[RegisterSet::val] = true; \ |
| if (preserved && (Include & ::Ice::TargetLowering::RegSet_CalleeSave)) \ |
| Registers[RegisterSet::val] = true; \ |
| if (stackptr && (Include & ::Ice::TargetLowering::RegSet_StackPointer)) \ |
| Registers[RegisterSet::val] = true; \ |
| if (frameptr && (Include & ::Ice::TargetLowering::RegSet_FramePointer)) \ |
| Registers[RegisterSet::val] = true; \ |
| if (scratch && (Exclude & ::Ice::TargetLowering::RegSet_CallerSave)) \ |
| Registers[RegisterSet::val] = false; \ |
| if (preserved && (Exclude & ::Ice::TargetLowering::RegSet_CalleeSave)) \ |
| Registers[RegisterSet::val] = false; \ |
| if (stackptr && (Exclude & ::Ice::TargetLowering::RegSet_StackPointer)) \ |
| Registers[RegisterSet::val] = false; \ |
| if (frameptr && (Exclude & ::Ice::TargetLowering::RegSet_FramePointer)) \ |
| Registers[RegisterSet::val] = false; \ |
| } |
| |
| REGX8664_TABLE |
| |
| #undef X |
| |
| return Registers; |
| } |
| |
| static void makeRandomRegisterPermutation( |
| Cfg *Func, llvm::SmallVectorImpl<RegNumT> &Permutation, |
| const SmallBitVector &ExcludeRegisters, uint64_t Salt) { |
| // TODO(stichnot): Declaring Permutation this way loses type/size |
| // information. Fix this in conjunction with the caller-side TODO. |
| assert(Permutation.size() >= RegisterSet::Reg_NUM); |
| // Expected upper bound on the number of registers in a single equivalence |
| // class. For x86-64, this would comprise the 16 XMM registers. This is |
| // for performance, not correctness. |
| static const unsigned MaxEquivalenceClassSize = 8; |
| using RegisterList = llvm::SmallVector<RegNumT, MaxEquivalenceClassSize>; |
| using EquivalenceClassMap = std::map<uint32_t, RegisterList>; |
| EquivalenceClassMap EquivalenceClasses; |
| SizeT NumShuffled = 0, NumPreserved = 0; |
| |
| // Build up the equivalence classes of registers by looking at the register |
| // properties as well as whether the registers should be explicitly excluded |
| // from shuffling. |
| #define X(val, encode, name, base, scratch, preserved, stackptr, frameptr, \ |
| sboxres, isGPR, is64, is32, is16, is8, isXmm, is64To8, is32To8, \ |
| is16To8, isTrunc8Rcvr, isAhRcvr, aliases) \ |
| if (ExcludeRegisters[RegisterSet::val]) { \ |
| /* val stays the same in the resulting permutation. */ \ |
| Permutation[RegisterSet::val] = RegisterSet::val; \ |
| ++NumPreserved; \ |
| } else { \ |
| uint32_t AttrKey = 0; \ |
| uint32_t Index = 0; \ |
| /* Combine relevant attributes into an equivalence class key. */ \ |
| Index |= (scratch << (AttrKey++)); \ |
| Index |= (preserved << (AttrKey++)); \ |
| Index |= (is8 << (AttrKey++)); \ |
| Index |= (is16 << (AttrKey++)); \ |
| Index |= (is32 << (AttrKey++)); \ |
| Index |= (is64 << (AttrKey++)); \ |
| Index |= (isXmm << (AttrKey++)); \ |
| Index |= (is16To8 << (AttrKey++)); \ |
| Index |= (is32To8 << (AttrKey++)); \ |
| Index |= (is64To8 << (AttrKey++)); \ |
| Index |= (isTrunc8Rcvr << (AttrKey++)); \ |
| /* val is assigned to an equivalence class based on its properties. */ \ |
| EquivalenceClasses[Index].push_back(RegisterSet::val); \ |
| } |
| REGX8664_TABLE |
| #undef X |
| |
| // Create a random number generator for regalloc randomization. |
| RandomNumberGenerator RNG(getFlags().getRandomSeed(), |
| RPE_RegAllocRandomization, Salt); |
| RandomNumberGeneratorWrapper RNGW(RNG); |
| |
| // Shuffle the resulting equivalence classes. |
| for (auto I : EquivalenceClasses) { |
| const RegisterList &List = I.second; |
| RegisterList Shuffled(List); |
| RandomShuffle(Shuffled.begin(), Shuffled.end(), RNGW); |
| for (size_t SI = 0, SE = Shuffled.size(); SI < SE; ++SI) { |
| Permutation[List[SI]] = Shuffled[SI]; |
| ++NumShuffled; |
| } |
| } |
| |
| assert(NumShuffled + NumPreserved == RegisterSet::Reg_NUM); |
| |
| if (Func->isVerbose(IceV_Random)) { |
| OstreamLocker L(Func->getContext()); |
| Ostream &Str = Func->getContext()->getStrDump(); |
| Str << "Register equivalence classes:\n"; |
| for (auto I : EquivalenceClasses) { |
| Str << "{"; |
| const RegisterList &List = I.second; |
| bool First = true; |
| for (RegNumT Register : List) { |
| if (!First) |
| Str << " "; |
| First = false; |
| Str << getRegName(Register); |
| } |
| Str << "}\n"; |
| } |
| } |
| } |
| |
| static RegNumT getRaxOrDie() { return RegisterSet::Reg_rax; } |
| |
| static RegNumT getRdxOrDie() { return RegisterSet::Reg_rdx; } |
| |
| #if defined(SUBZERO_USE_MICROSOFT_ABI) |
| // Microsoft x86-64 calling convention: |
| // |
| // * The first four arguments of vector/fp type, regardless of their |
| // position relative to the other arguments in the argument list, are placed |
| // in registers %xmm0 - %xmm3. |
| // |
| // * The first four arguments of integer types, regardless of their position |
| // relative to the other arguments in the argument list, are placed in |
| // registers %rcx, %rdx, %r8, and %r9. |
| |
| /// The maximum number of arguments to pass in XMM registers |
| static constexpr uint32_t X86_MAX_XMM_ARGS = 4; |
| /// The maximum number of arguments to pass in GPR registers |
| static constexpr uint32_t X86_MAX_GPR_ARGS = 4; |
| static RegNumT getRegisterForGprArgNum(Type Ty, uint32_t ArgNum) { |
| if (ArgNum >= X86_MAX_GPR_ARGS) { |
| return RegNumT(); |
| } |
| static const RegisterSet::AllRegisters GprForArgNum[] = { |
| RegisterSet::Reg_rcx, RegisterSet::Reg_rdx, RegisterSet::Reg_r8, |
| RegisterSet::Reg_r9, |
| }; |
| static_assert(llvm::array_lengthof(GprForArgNum) == X86_MAX_GPR_ARGS, |
| "Mismatch between MAX_GPR_ARGS and GprForArgNum."); |
| assert(Ty == IceType_i64 || Ty == IceType_i32); |
| return getGprForType(Ty, GprForArgNum[ArgNum]); |
| } |
| #else |
| // System V x86-64 calling convention: |
| // |
| // * The first eight arguments of vector/fp type, regardless of their |
| // position relative to the other arguments in the argument list, are placed |
| // in registers %xmm0 - %xmm7. |
| // |
| // * The first six arguments of integer types, regardless of their position |
| // relative to the other arguments in the argument list, are placed in |
| // registers %rdi, %rsi, %rdx, %rcx, %r8, and %r9. |
| // |
| // This intends to match the section "Function Calling Sequence" of the |
| // document "System V Application Binary Interface." |
| |
| /// The maximum number of arguments to pass in XMM registers |
| static constexpr uint32_t X86_MAX_XMM_ARGS = 8; |
| /// The maximum number of arguments to pass in GPR registers |
| static constexpr uint32_t X86_MAX_GPR_ARGS = 6; |
| /// Get the register for a given argument slot in the GPRs. |
| static RegNumT getRegisterForGprArgNum(Type Ty, uint32_t ArgNum) { |
| if (ArgNum >= X86_MAX_GPR_ARGS) { |
| return RegNumT(); |
| } |
| static const RegisterSet::AllRegisters GprForArgNum[] = { |
| RegisterSet::Reg_rdi, RegisterSet::Reg_rsi, RegisterSet::Reg_rdx, |
| RegisterSet::Reg_rcx, RegisterSet::Reg_r8, RegisterSet::Reg_r9, |
| }; |
| static_assert(llvm::array_lengthof(GprForArgNum) == X86_MAX_GPR_ARGS, |
| "Mismatch between MAX_GPR_ARGS and GprForArgNum."); |
| assert(Ty == IceType_i64 || Ty == IceType_i32); |
| return getGprForType(Ty, GprForArgNum[ArgNum]); |
| } |
| #endif |
| |
| /// Whether scalar floating point arguments are passed in XMM registers |
| static constexpr bool X86_PASS_SCALAR_FP_IN_XMM = true; |
| /// Get the register for a given argument slot in the XMM registers. |
| static RegNumT getRegisterForXmmArgNum(uint32_t ArgNum) { |
| // TODO(sehr): Change to use the CCArg technique used in ARM32. |
| static_assert(RegisterSet::Reg_xmm0 + 1 == RegisterSet::Reg_xmm1, |
| "Inconsistency between XMM register numbers and ordinals"); |
| if (ArgNum >= X86_MAX_XMM_ARGS) { |
| return RegNumT(); |
| } |
| return RegNumT::fixme(RegisterSet::Reg_xmm0 + ArgNum); |
| } |
| |
| /// The number of bits in a byte |
| static constexpr uint32_t X86_CHAR_BIT = 8; |
| /// Stack alignment. This is defined in IceTargetLoweringX8664.cpp because it |
| /// is used as an argument to std::max(), and the default std::less<T> has an |
| /// operator(T const&, T const&) which requires this member to have an |
| /// address. |
| static const uint32_t X86_STACK_ALIGNMENT_BYTES; |
| /// Size of the return address on the stack |
| static constexpr uint32_t X86_RET_IP_SIZE_BYTES = 8; |
| /// The number of different NOP instructions |
| static constexpr uint32_t X86_NUM_NOP_VARIANTS = 5; |
| |
| /// \name Limits for unrolling memory intrinsics. |
| /// @{ |
| static constexpr uint32_t MEMCPY_UNROLL_LIMIT = 8; |
| static constexpr uint32_t MEMMOVE_UNROLL_LIMIT = 8; |
| static constexpr uint32_t MEMSET_UNROLL_LIMIT = 8; |
| /// @} |
| |
| /// Value is in bytes. Return Value adjusted to the next highest multiple of |
| /// the stack alignment. |
| static uint32_t applyStackAlignment(uint32_t Value) { |
| return Utils::applyAlignment(Value, X86_STACK_ALIGNMENT_BYTES); |
| } |
| |
| /// Return the type which the elements of the vector have in the X86 |
| /// representation of the vector. |
| static Type getInVectorElementType(Type Ty) { |
| assert(isVectorType(Ty)); |
| assert(Ty < TableTypeX8664AttributesSize); |
| return TableTypeX8664Attributes[Ty].InVectorElementType; |
| } |
| |
| // Note: The following data structures are defined in |
| // IceTargetLoweringX8664.cpp. |
| |
| /// The following table summarizes the logic for lowering the fcmp |
| /// instruction. There is one table entry for each of the 16 conditions. |
| /// |
| /// The first four columns describe the case when the operands are floating |
| /// point scalar values. A comment in lowerFcmp() describes the lowering |
| /// template. In the most general case, there is a compare followed by two |
| /// conditional branches, because some fcmp conditions don't map to a single |
| /// x86 conditional branch. However, in many cases it is possible to swap the |
| /// operands in the comparison and have a single conditional branch. Since |
| /// it's quite tedious to validate the table by hand, good execution tests are |
| /// helpful. |
| /// |
| /// The last two columns describe the case when the operands are vectors of |
| /// floating point values. For most fcmp conditions, there is a clear mapping |
| /// to a single x86 cmpps instruction variant. Some fcmp conditions require |
| /// special code to handle and these are marked in the table with a |
| /// Cmpps_Invalid predicate. |
| /// {@ |
| static const struct TableFcmpType { |
| uint32_t Default; |
| bool SwapScalarOperands; |
| Cond::BrCond C1, C2; |
| bool SwapVectorOperands; |
| Cond::CmppsCond Predicate; |
| } TableFcmp[]; |
| static const size_t TableFcmpSize; |
| /// @} |
| |
| /// The following table summarizes the logic for lowering the icmp instruction |
| /// for i32 and narrower types. Each icmp condition has a clear mapping to an |
| /// x86 conditional branch instruction. |
| /// {@ |
| static const struct TableIcmp32Type { Cond::BrCond Mapping; } TableIcmp32[]; |
| static const size_t TableIcmp32Size; |
| /// @} |
| |
| /// The following table summarizes the logic for lowering the icmp instruction |
| /// for the i64 type. For Eq and Ne, two separate 32-bit comparisons and |
| /// conditional branches are needed. For the other conditions, three separate |
| /// conditional branches are needed. |
| /// {@ |
| static const struct TableIcmp64Type { |
| Cond::BrCond C1, C2, C3; |
| } TableIcmp64[]; |
| static const size_t TableIcmp64Size; |
| /// @} |
| |
| static Cond::BrCond getIcmp32Mapping(InstIcmp::ICond Cond) { |
| assert(Cond < TableIcmp32Size); |
| return TableIcmp32[Cond].Mapping; |
| } |
| |
| static const struct TableTypeX8664AttributesType { |
| Type InVectorElementType; |
| } TableTypeX8664Attributes[]; |
| static const size_t TableTypeX8664AttributesSize; |
| |
| //---------------------------------------------------------------------------- |
| // __ __ __ ______ ______ |
| // /\ \/\ "-.\ \/\ ___\/\__ _\ |
| // \ \ \ \ \-. \ \___ \/_/\ \/ |
| // \ \_\ \_\\"\_\/\_____\ \ \_\ |
| // \/_/\/_/ \/_/\/_____/ \/_/ |
| // |
| //---------------------------------------------------------------------------- |
| using Traits = TargetX8664Traits; |
| using Insts = ::Ice::X8664::Insts<Traits>; |
| |
| using TargetLowering = ::Ice::X8664::TargetX86Base<Traits>; |
| using ConcreteTarget = ::Ice::X8664::TargetX8664; |
| using Assembler = ::Ice::X8664::AssemblerX86Base<Traits>; |
| |
| /// X86Operand extends the Operand hierarchy. Its subclasses are X86OperandMem |
| /// and VariableSplit. |
| class X86Operand : public ::Ice::Operand { |
| X86Operand() = delete; |
| X86Operand(const X86Operand &) = delete; |
| X86Operand &operator=(const X86Operand &) = delete; |
| |
| public: |
| enum OperandKindX8664 { k__Start = ::Ice::Operand::kTarget, kMem, kSplit }; |
| using ::Ice::Operand::dump; |
| |
| void dump(const Cfg *, Ostream &Str) const override; |
| |
| protected: |
| X86Operand(OperandKindX8664 Kind, Type Ty) |
| : Operand(static_cast<::Ice::Operand::OperandKind>(Kind), Ty) {} |
| }; |
| |
| /// X86OperandMem represents the m64 addressing mode, with optional base and |
| /// index registers, a constant offset, and a fixed shift value for the index |
| /// register. |
| class X86OperandMem : public X86Operand { |
| X86OperandMem() = delete; |
| X86OperandMem(const X86OperandMem &) = delete; |
| X86OperandMem &operator=(const X86OperandMem &) = delete; |
| |
| public: |
| enum SegmentRegisters { DefaultSegment = -1, SegReg_NUM }; |
| static X86OperandMem * |
| create(Cfg *Func, Type Ty, Variable *Base, Constant *Offset, |
| Variable *Index = nullptr, uint16_t Shift = 0, |
| SegmentRegisters SegmentRegister = DefaultSegment, |
| bool IsRebased = false) { |
| assert(SegmentRegister == DefaultSegment); |
| (void)SegmentRegister; |
| return new (Func->allocate<X86OperandMem>()) |
| X86OperandMem(Func, Ty, Base, Offset, Index, Shift, IsRebased); |
| } |
| static X86OperandMem *create(Cfg *Func, Type Ty, Variable *Base, |
| Constant *Offset, bool IsRebased) { |
| constexpr Variable *NoIndex = nullptr; |
| constexpr uint16_t NoShift = 0; |
| return new (Func->allocate<X86OperandMem>()) |
| X86OperandMem(Func, Ty, Base, Offset, NoIndex, NoShift, IsRebased); |
| } |
| Variable *getBase() const { return Base; } |
| Constant *getOffset() const { return Offset; } |
| Variable *getIndex() const { return Index; } |
| uint16_t getShift() const { return Shift; } |
| SegmentRegisters getSegmentRegister() const { return DefaultSegment; } |
| void emitSegmentOverride(Assembler *) const {} |
| bool getIsRebased() const { return IsRebased; } |
| Address toAsmAddress(Assembler *Asm, const Ice::TargetLowering *Target, |
| bool IsLeaAddr = false) const; |
| |
| void emit(const Cfg *Func) const override; |
| using X86Operand::dump; |
| void dump(const Cfg *Func, Ostream &Str) const override; |
| |
| static bool classof(const Operand *Operand) { |
| return Operand->getKind() == static_cast<OperandKind>(kMem); |
| } |
| |
| void setRandomized(bool R) { Randomized = R; } |
| |
| bool getRandomized() const { return Randomized; } |
| |
| private: |
| X86OperandMem(Cfg *Func, Type Ty, Variable *Base, Constant *Offset, |
| Variable *Index, uint16_t Shift, bool IsRebased); |
| |
| Variable *const Base; |
| Constant *const Offset; |
| Variable *const Index; |
| const uint16_t Shift; |
| const bool IsRebased; |
| /// A flag to show if this memory operand is a randomized one. Randomized |
| /// memory operands are generated in |
| /// TargetX86Base::randomizeOrPoolImmediate() |
| bool Randomized = false; |
| }; |
| |
| /// VariableSplit is a way to treat an f64 memory location as a pair of i32 |
| /// locations (Low and High). This is needed for some cases of the Bitcast |
| /// instruction. Since it's not possible for integer registers to access the |
| /// XMM registers and vice versa, the lowering forces the f64 to be spilled to |
| /// the stack and then accesses through the VariableSplit. |
| // TODO(jpp): remove references to VariableSplit from IceInstX86Base as 64bit |
| // targets can natively handle these. |
| class VariableSplit : public X86Operand { |
| VariableSplit() = delete; |
| VariableSplit(const VariableSplit &) = delete; |
| VariableSplit &operator=(const VariableSplit &) = delete; |
| |
| public: |
| enum Portion { Low, High }; |
| static VariableSplit *create(Cfg *Func, Variable *Var, Portion Part) { |
| return new (Func->allocate<VariableSplit>()) |
| VariableSplit(Func, Var, Part); |
| } |
| int32_t getOffset() const { return Part == High ? 4 : 0; } |
| |
| Address toAsmAddress(const Cfg *Func) const; |
| void emit(const Cfg *Func) const override; |
| using X86Operand::dump; |
| void dump(const Cfg *Func, Ostream &Str) const override; |
| |
| static bool classof(const Operand *Operand) { |
| return Operand->getKind() == static_cast<OperandKind>(kSplit); |
| } |
| |
| private: |
| VariableSplit(Cfg *Func, Variable *Var, Portion Part) |
| : X86Operand(kSplit, IceType_i32), Var(Var), Part(Part) { |
| assert(Var->getType() == IceType_f64); |
| Vars = Func->allocateArrayOf<Variable *>(1); |
| Vars[0] = Var; |
| NumVars = 1; |
| } |
| |
| Variable *Var; |
| Portion Part; |
| }; |
| |
| // Note: The following data structures are defined in IceInstX8664.cpp. |
| |
| static const struct InstBrAttributesType { |
| Cond::BrCond Opposite; |
| const char *DisplayString; |
| const char *EmitString; |
| } InstBrAttributes[]; |
| |
| static const struct InstCmppsAttributesType { |
| const char *EmitString; |
| } InstCmppsAttributes[]; |
| |
| static const struct TypeAttributesType { |
| const char *CvtString; // i (integer), s (single FP), d (double FP) |
| const char *SdSsString; // ss, sd, or <blank> |
| const char *PdPsString; // ps, pd, or <blank> |
| const char *SpSdString; // ss, sd, ps, pd, or <blank> |
| const char *IntegralString; // b, w, d, or <blank> |
| const char *UnpackString; // bw, wd, dq, or <blank> |
| const char *PackString; // wb, dw, or <blank> |
| const char *WidthString; // b, w, l, q, or <blank> |
| const char *FldString; // s, l, or <blank> |
| } TypeAttributes[]; |
| }; |
| |
| using Traits = ::Ice::X8664::TargetX8664Traits; |
| } // end of namespace X8664 |
| |
| } // end of namespace Ice |
| |
| #endif // SUBZERO_SRC_ICETARGETLOWERINGX8664TRAITS_H |