|  | //===- subzero/src/IceTargetLoweringMIPS32.h - MIPS32 lowering ---*- C++-*-===// | 
|  | // | 
|  | //                        The Subzero Code Generator | 
|  | // | 
|  | // This file is distributed under the University of Illinois Open Source | 
|  | // License. See LICENSE.TXT for details. | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  | /// | 
|  | /// \file | 
|  | /// \brief Declares the TargetLoweringMIPS32 class, which implements the | 
|  | /// TargetLowering interface for the MIPS 32-bit architecture. | 
|  | /// | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #ifndef SUBZERO_SRC_ICETARGETLOWERINGMIPS32_H | 
|  | #define SUBZERO_SRC_ICETARGETLOWERINGMIPS32_H | 
|  |  | 
|  | #include "IceAssemblerMIPS32.h" | 
|  | #include "IceDefs.h" | 
|  | #include "IceInstMIPS32.h" | 
|  | #include "IceRegistersMIPS32.h" | 
|  | #include "IceTargetLowering.h" | 
|  |  | 
|  | namespace Ice { | 
|  | namespace MIPS32 { | 
|  |  | 
|  | class TargetMIPS32 : public TargetLowering { | 
|  | TargetMIPS32() = delete; | 
|  | TargetMIPS32(const TargetMIPS32 &) = delete; | 
|  | TargetMIPS32 &operator=(const TargetMIPS32 &) = delete; | 
|  |  | 
|  | public: | 
|  | ~TargetMIPS32() override = default; | 
|  |  | 
|  | static void staticInit(GlobalContext *Ctx); | 
|  | static bool shouldBePooled(const Constant *C) { | 
|  | if (auto *ConstDouble = llvm::dyn_cast<ConstantDouble>(C)) { | 
|  | return !Utils::isPositiveZero(ConstDouble->getValue()); | 
|  | } | 
|  | if (auto *ConstFloat = llvm::dyn_cast<ConstantFloat>(C)) { | 
|  | return !Utils::isPositiveZero(ConstFloat->getValue()); | 
|  | } | 
|  | return false; | 
|  | } | 
|  | static ::Ice::Type getPointerType() { return ::Ice::IceType_i32; } | 
|  | static std::unique_ptr<::Ice::TargetLowering> create(Cfg *Func) { | 
|  | return makeUnique<TargetMIPS32>(Func); | 
|  | } | 
|  |  | 
|  | std::unique_ptr<::Ice::Assembler> createAssembler() const override { | 
|  | return makeUnique<MIPS32::AssemblerMIPS32>(); | 
|  | } | 
|  |  | 
|  | void initNodeForLowering(CfgNode *Node) override { | 
|  | Computations.forgetProducers(); | 
|  | Computations.recordProducers(Node); | 
|  | Computations.dump(Func); | 
|  | } | 
|  |  | 
|  | void translateOm1() override; | 
|  | void translateO2() override; | 
|  | bool doBranchOpt(Inst *Instr, const CfgNode *NextNode) override; | 
|  | void setImplicitRet(Variable *Ret) { ImplicitRet = Ret; } | 
|  | Variable *getImplicitRet() const { return ImplicitRet; } | 
|  | SizeT getNumRegisters() const override { return RegMIPS32::Reg_NUM; } | 
|  | Variable *getPhysicalRegister(RegNumT RegNum, | 
|  | Type Ty = IceType_void) override; | 
|  | const char *getRegName(RegNumT RegNum, Type Ty) const override; | 
|  | SmallBitVector getRegisterSet(RegSetMask Include, | 
|  | RegSetMask Exclude) const override; | 
|  | const SmallBitVector & | 
|  | getRegistersForVariable(const Variable *Var) const override { | 
|  | RegClass RC = Var->getRegClass(); | 
|  | assert(RC < RC_Target); | 
|  | return TypeToRegisterSet[RC]; | 
|  | } | 
|  | const SmallBitVector & | 
|  | getAllRegistersForVariable(const Variable *Var) const override { | 
|  | RegClass RC = Var->getRegClass(); | 
|  | assert(RC < RC_Target); | 
|  | return TypeToRegisterSetUnfiltered[RC]; | 
|  | } | 
|  | const SmallBitVector &getAliasesForRegister(RegNumT Reg) const override { | 
|  | return RegisterAliases[Reg]; | 
|  | } | 
|  | bool hasFramePointer() const override { return UsesFramePointer; } | 
|  | void setHasFramePointer() override { UsesFramePointer = true; } | 
|  | RegNumT getStackReg() const override { return RegMIPS32::Reg_SP; } | 
|  | RegNumT getFrameReg() const override { return RegMIPS32::Reg_FP; } | 
|  | RegNumT getFrameOrStackReg() const override { | 
|  | return UsesFramePointer ? getFrameReg() : getStackReg(); | 
|  | } | 
|  | RegNumT getReservedTmpReg() const { return RegMIPS32::Reg_AT; } | 
|  | size_t typeWidthInBytesOnStack(Type Ty) const override { | 
|  | // Round up to the next multiple of 4 bytes. In particular, i1, i8, and i16 | 
|  | // are rounded up to 4 bytes. | 
|  | return (typeWidthInBytes(Ty) + 3) & ~3; | 
|  | } | 
|  | uint32_t getStackAlignment() const override; | 
|  | void reserveFixedAllocaArea(size_t Size, size_t Align) override { | 
|  | FixedAllocaSizeBytes = Size; | 
|  | assert(llvm::isPowerOf2_32(Align)); | 
|  | FixedAllocaAlignBytes = Align; | 
|  | PrologEmitsFixedAllocas = true; | 
|  | } | 
|  | int32_t getFrameFixedAllocaOffset() const override { | 
|  | int32_t FixedAllocaOffset = | 
|  | Utils::applyAlignment(CurrentAllocaOffset, FixedAllocaAlignBytes); | 
|  | return FixedAllocaOffset - MaxOutArgsSizeBytes; | 
|  | } | 
|  |  | 
|  | uint32_t maxOutArgsSizeBytes() const override { return MaxOutArgsSizeBytes; } | 
|  |  | 
|  | uint32_t getFramePointerOffset(uint32_t CurrentOffset, | 
|  | uint32_t Size) const override { | 
|  | (void)Size; | 
|  | return CurrentOffset + MaxOutArgsSizeBytes; | 
|  | } | 
|  |  | 
|  | bool shouldSplitToVariable64On32(Type Ty) const override { | 
|  | return Ty == IceType_i64; | 
|  | } | 
|  |  | 
|  | bool shouldSplitToVariableVecOn32(Type Ty) const override { | 
|  | return isVectorType(Ty); | 
|  | } | 
|  |  | 
|  | // TODO(ascull): what is the best size of MIPS? | 
|  | SizeT getMinJumpTableSize() const override { return 3; } | 
|  | void emitJumpTable(const Cfg *Func, | 
|  | const InstJumpTable *JumpTable) const override; | 
|  |  | 
|  | void emitVariable(const Variable *Var) const override; | 
|  |  | 
|  | void emit(const ConstantInteger32 *C) const final { | 
|  | if (!BuildDefs::dump()) | 
|  | return; | 
|  | Ostream &Str = Ctx->getStrEmit(); | 
|  | Str << C->getValue(); | 
|  | } | 
|  | void emit(const ConstantInteger64 *C) const final { | 
|  | (void)C; | 
|  | llvm::report_fatal_error("Not yet implemented"); | 
|  | } | 
|  | void emit(const ConstantFloat *C) const final { | 
|  | (void)C; | 
|  | llvm::report_fatal_error("Not yet implemented"); | 
|  | } | 
|  | void emit(const ConstantDouble *C) const final { | 
|  | (void)C; | 
|  | llvm::report_fatal_error("Not yet implemented"); | 
|  | } | 
|  | void emit(const ConstantUndef *C) const final { | 
|  | (void)C; | 
|  | llvm::report_fatal_error("Not yet implemented"); | 
|  | } | 
|  | void emit(const ConstantRelocatable *C) const final { | 
|  | (void)C; | 
|  | llvm::report_fatal_error("Not yet implemented"); | 
|  | } | 
|  |  | 
|  | // The following are helpers that insert lowered MIPS32 instructions with | 
|  | // minimal syntactic overhead, so that the lowering code can look as close to | 
|  | // assembly as practical. | 
|  | void _add(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Add>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _addu(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Addu>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _and(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32And>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _andi(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Andi>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _br(CfgNode *Target) { Context.insert<InstMIPS32Br>(Target); } | 
|  |  | 
|  | void _br(CfgNode *Target, const InstMIPS32Label *Label) { | 
|  | Context.insert<InstMIPS32Br>(Target, Label); | 
|  | } | 
|  |  | 
|  | void _br(CfgNode *TargetTrue, CfgNode *TargetFalse, Operand *Src0, | 
|  | Operand *Src1, CondMIPS32::Cond Condition) { | 
|  | Context.insert<InstMIPS32Br>(TargetTrue, TargetFalse, Src0, Src1, | 
|  | Condition); | 
|  | } | 
|  |  | 
|  | void _br(CfgNode *TargetTrue, CfgNode *TargetFalse, Operand *Src0, | 
|  | CondMIPS32::Cond Condition) { | 
|  | Context.insert<InstMIPS32Br>(TargetTrue, TargetFalse, Src0, Condition); | 
|  | } | 
|  |  | 
|  | void _br(CfgNode *TargetTrue, CfgNode *TargetFalse, Operand *Src0, | 
|  | Operand *Src1, const InstMIPS32Label *Label, | 
|  | CondMIPS32::Cond Condition) { | 
|  | Context.insert<InstMIPS32Br>(TargetTrue, TargetFalse, Src0, Src1, Label, | 
|  | Condition); | 
|  | } | 
|  |  | 
|  | void _ret(Variable *RA, Variable *Src0 = nullptr) { | 
|  | Context.insert<InstMIPS32Ret>(RA, Src0); | 
|  | } | 
|  |  | 
|  | void _abs_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Abs_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _abs_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Abs_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _addi(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Addi>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _add_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Add_d>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _add_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Add_s>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _addiu(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Addiu>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _addiu(Variable *Dest, Variable *Src0, Operand *Src1, RelocOp Reloc) { | 
|  | Context.insert<InstMIPS32Addiu>(Dest, Src0, Src1, Reloc); | 
|  | } | 
|  |  | 
|  | void _c_eq_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_eq_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_eq_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_eq_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ole_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ole_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ole_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ole_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_olt_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_olt_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_olt_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_olt_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ueq_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ueq_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ueq_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ueq_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ule_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ule_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ule_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ule_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ult_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ult_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_ult_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_ult_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_un_d(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_un_d>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _c_un_s(Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32C_un_s>(Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _clz(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Clz>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_d_l(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_d_l>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_d_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_d_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_d_w(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_d_w>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_s_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_s_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_s_l(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_s_l>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _cvt_s_w(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Cvt_s_w>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _div(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Div>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _div_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Div_d>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _div_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Div_s>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _divu(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Divu>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _ldc1(Variable *Value, OperandMIPS32Mem *Mem, RelocOp Reloc = RO_No) { | 
|  | Context.insert<InstMIPS32Ldc1>(Value, Mem, Reloc); | 
|  | } | 
|  |  | 
|  | void _ll(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | Context.insert<InstMIPS32Ll>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _lw(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | Context.insert<InstMIPS32Lw>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _lwc1(Variable *Value, OperandMIPS32Mem *Mem, RelocOp Reloc = RO_No) { | 
|  | Context.insert<InstMIPS32Lwc1>(Value, Mem, Reloc); | 
|  | } | 
|  |  | 
|  | void _lui(Variable *Dest, Operand *Src, RelocOp Reloc = RO_No) { | 
|  | Context.insert<InstMIPS32Lui>(Dest, Src, Reloc); | 
|  | } | 
|  |  | 
|  | void _mfc1(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Mfc1>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mfhi(Variable *Dest, Operand *Src) { | 
|  | Context.insert<InstMIPS32Mfhi>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mflo(Variable *Dest, Operand *Src) { | 
|  | Context.insert<InstMIPS32Mflo>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mov(Variable *Dest, Operand *Src0, Operand *Src1 = nullptr) { | 
|  | assert(Dest != nullptr); | 
|  | // Variable* Src0_ = llvm::dyn_cast<Variable>(Src0); | 
|  | if (llvm::isa<ConstantRelocatable>(Src0)) { | 
|  | Context.insert<InstMIPS32La>(Dest, Src0); | 
|  | } else { | 
|  | auto *Instr = Context.insert<InstMIPS32Mov>(Dest, Src0, Src1); | 
|  | if (Instr->getDestHi() != nullptr) { | 
|  | // If DestHi is available, then Dest must be a Variable64On32. We add a | 
|  | // fake-def for Instr.DestHi here. | 
|  | assert(llvm::isa<Variable64On32>(Dest)); | 
|  | Context.insert<InstFakeDef>(Instr->getDestHi()); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void _mov_redefined(Variable *Dest, Operand *Src0, Operand *Src1 = nullptr) { | 
|  | if (llvm::isa<ConstantRelocatable>(Src0)) { | 
|  | Context.insert<InstMIPS32La>(Dest, Src0); | 
|  | } else { | 
|  | auto *Instr = Context.insert<InstMIPS32Mov>(Dest, Src0, Src1); | 
|  | Instr->setDestRedefined(); | 
|  | if (Instr->getDestHi() != nullptr) { | 
|  | // If Instr is multi-dest, then Dest must be a Variable64On32. We add a | 
|  | // fake-def for Instr.DestHi here. | 
|  | assert(llvm::isa<Variable64On32>(Dest)); | 
|  | Context.insert<InstFakeDef>(Instr->getDestHi()); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void _mov_fp64_to_i64(Variable *Dest, Operand *Src, Int64Part Int64HiLo) { | 
|  | assert(Dest != nullptr); | 
|  | Context.insert<InstMIPS32MovFP64ToI64>(Dest, Src, Int64HiLo); | 
|  | } | 
|  |  | 
|  | void _mov_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Mov_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mov_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Mov_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _movf(Variable *Dest, Variable *Src0, Operand *FCC) { | 
|  | Context.insert<InstMIPS32Movf>(Dest, Src0, FCC)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movn(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movn>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movn_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movn_d>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movn_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movn_s>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movt(Variable *Dest, Variable *Src0, Operand *FCC) { | 
|  | Context.insert<InstMIPS32Movt>(Dest, Src0, FCC)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movz(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movz>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movz_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movz_d>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _movz_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Movz_s>(Dest, Src0, Src1)->setDestRedefined(); | 
|  | } | 
|  |  | 
|  | void _mtc1(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Mtc1>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mthi(Variable *Dest, Operand *Src) { | 
|  | Context.insert<InstMIPS32Mthi>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mtlo(Variable *Dest, Operand *Src) { | 
|  | Context.insert<InstMIPS32Mtlo>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _mul(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Mul>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _mul_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Mul_d>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _mul_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Mul_s>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _mult(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Mult>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _multu(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Multu>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _nop() { Context.insert<InstMIPS32Sll>(getZero(), getZero(), 0); } | 
|  |  | 
|  | void _nor(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Nor>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _not(Variable *Dest, Variable *Src0) { | 
|  | Context.insert<InstMIPS32Nor>(Dest, Src0, getZero()); | 
|  | } | 
|  |  | 
|  | void _or(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Or>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _ori(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Ori>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | InstMIPS32Sc *_sc(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | return Context.insert<InstMIPS32Sc>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _sdc1(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | Context.insert<InstMIPS32Sdc1>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _sll(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Sll>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _sllv(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Sllv>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _slt(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Slt>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _slti(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Slti>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _sltiu(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Sltiu>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _sltu(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Sltu>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _sqrt_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Sqrt_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _sqrt_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Sqrt_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _sra(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Sra>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _srav(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Srav>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _srl(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Srl>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void _srlv(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Srlv>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _sub(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Sub>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _sub_d(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Sub_d>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _sub_s(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Sub_s>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _subu(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Subu>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _sw(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | Context.insert<InstMIPS32Sw>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _swc1(Variable *Value, OperandMIPS32Mem *Mem) { | 
|  | Context.insert<InstMIPS32Swc1>(Value, Mem); | 
|  | } | 
|  |  | 
|  | void _sync() { Context.insert<InstMIPS32Sync>(); } | 
|  |  | 
|  | void _teq(Variable *Src0, Variable *Src1, uint32_t TrapCode) { | 
|  | Context.insert<InstMIPS32Teq>(Src0, Src1, TrapCode); | 
|  | } | 
|  |  | 
|  | void _trunc_l_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Trunc_l_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _trunc_l_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Trunc_l_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _trunc_w_d(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Trunc_w_d>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _trunc_w_s(Variable *Dest, Variable *Src) { | 
|  | Context.insert<InstMIPS32Trunc_w_s>(Dest, Src); | 
|  | } | 
|  |  | 
|  | void _xor(Variable *Dest, Variable *Src0, Variable *Src1) { | 
|  | Context.insert<InstMIPS32Xor>(Dest, Src0, Src1); | 
|  | } | 
|  |  | 
|  | void _xori(Variable *Dest, Variable *Src, uint32_t Imm) { | 
|  | Context.insert<InstMIPS32Xori>(Dest, Src, Imm); | 
|  | } | 
|  |  | 
|  | void lowerArguments() override; | 
|  |  | 
|  | /// Make a pass through the SortedSpilledVariables and actually assign stack | 
|  | /// slots. SpillAreaPaddingBytes takes into account stack alignment padding. | 
|  | /// The SpillArea starts after that amount of padding. This matches the scheme | 
|  | /// in getVarStackSlotParams, where there may be a separate multi-block global | 
|  | /// var spill area and a local var spill area. | 
|  | void assignVarStackSlots(VarList &SortedSpilledVariables, | 
|  | size_t SpillAreaPaddingBytes, | 
|  | size_t SpillAreaSizeBytes, | 
|  | size_t GlobalsAndSubsequentPaddingSize); | 
|  |  | 
|  | /// Operand legalization helpers.  To deal with address mode constraints, | 
|  | /// the helpers will create a new Operand and emit instructions that | 
|  | /// guarantee that the Operand kind is one of those indicated by the | 
|  | /// LegalMask (a bitmask of allowed kinds).  If the input Operand is known | 
|  | /// to already meet the constraints, it may be simply returned as the result, | 
|  | /// without creating any new instructions or operands. | 
|  | enum OperandLegalization { | 
|  | Legal_None = 0, | 
|  | Legal_Reg = 1 << 0, // physical register, not stack location | 
|  | Legal_Imm = 1 << 1, | 
|  | Legal_Mem = 1 << 2, | 
|  | Legal_Rematerializable = 1 << 3, | 
|  | Legal_Default = ~Legal_None | 
|  | }; | 
|  | typedef uint32_t LegalMask; | 
|  | Operand *legalize(Operand *From, LegalMask Allowed = Legal_Default, | 
|  | RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | Variable *legalizeToVar(Operand *From, RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | Variable *legalizeToReg(Operand *From, RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | Variable *makeReg(Type Ty, RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | Variable *getZero() { | 
|  | auto *Zero = makeReg(IceType_i32, RegMIPS32::Reg_ZERO); | 
|  | Context.insert<InstFakeDef>(Zero); | 
|  | return Zero; | 
|  | } | 
|  |  | 
|  | Variable *I32Reg(RegNumT RegNum = RegNumT()) { | 
|  | return makeReg(IceType_i32, RegNum); | 
|  | } | 
|  |  | 
|  | Variable *F32Reg(RegNumT RegNum = RegNumT()) { | 
|  | return makeReg(IceType_f32, RegNum); | 
|  | } | 
|  |  | 
|  | Variable *F64Reg(RegNumT RegNum = RegNumT()) { | 
|  | return makeReg(IceType_f64, RegNum); | 
|  | } | 
|  |  | 
|  | static Type stackSlotType(); | 
|  | Variable *copyToReg(Operand *Src, RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | void unsetIfNonLeafFunc(); | 
|  |  | 
|  | // Iterates over the CFG and determines the maximum outgoing stack arguments | 
|  | // bytes. This information is later used during addProlog() to pre-allocate | 
|  | // the outargs area | 
|  | void findMaxStackOutArgsSize(); | 
|  |  | 
|  | void postLowerLegalization(); | 
|  |  | 
|  | void addProlog(CfgNode *Node) override; | 
|  | void addEpilog(CfgNode *Node) override; | 
|  |  | 
|  | // Ensure that a 64-bit Variable has been split into 2 32-bit | 
|  | // Variables, creating them if necessary.  This is needed for all | 
|  | // I64 operations. | 
|  | void split64(Variable *Var); | 
|  | Operand *loOperand(Operand *Operand); | 
|  | Operand *hiOperand(Operand *Operand); | 
|  | Operand *getOperandAtIndex(Operand *Operand, Type BaseType, uint32_t Index); | 
|  |  | 
|  | void finishArgumentLowering(Variable *Arg, bool PartialOnStack, | 
|  | Variable *FramePtr, size_t BasicFrameOffset, | 
|  | size_t *InArgsSizeBytes); | 
|  |  | 
|  | Operand *legalizeUndef(Operand *From, RegNumT RegNum = RegNumT()); | 
|  |  | 
|  | /// Helper class that understands the Calling Convention and register | 
|  | /// assignments as per MIPS O32 abi. | 
|  | class CallingConv { | 
|  | CallingConv(const CallingConv &) = delete; | 
|  | CallingConv &operator=(const CallingConv &) = delete; | 
|  |  | 
|  | public: | 
|  | CallingConv(); | 
|  | ~CallingConv() = default; | 
|  |  | 
|  | /// argInReg returns true if there is a Register available for the requested | 
|  | /// type, and false otherwise. If it returns true, Reg is set to the | 
|  | /// appropriate register number. Note that, when Ty == IceType_i64, Reg will | 
|  | /// be an I64 register pair. | 
|  | bool argInReg(Type Ty, uint32_t ArgNo, RegNumT *Reg); | 
|  | void discardReg(RegNumT Reg) { GPRegsUsed |= RegisterAliases[Reg]; } | 
|  |  | 
|  | private: | 
|  | // argInGPR is used to find if any GPR register is available for argument of | 
|  | // type Ty | 
|  | bool argInGPR(Type Ty, RegNumT *Reg); | 
|  | /// argInVFP is to floating-point/vector types what argInGPR is for integer | 
|  | /// types. | 
|  | bool argInVFP(Type Ty, RegNumT *Reg); | 
|  | inline void discardNextGPRAndItsAliases(CfgVector<RegNumT> *Regs); | 
|  | inline void alignGPR(CfgVector<RegNumT> *Regs); | 
|  | void discardUnavailableGPRsAndTheirAliases(CfgVector<RegNumT> *Regs); | 
|  | SmallBitVector GPRegsUsed; | 
|  | CfgVector<RegNumT> GPRArgs; | 
|  | CfgVector<RegNumT> I64Args; | 
|  |  | 
|  | void discardUnavailableVFPRegsAndTheirAliases(CfgVector<RegNumT> *Regs); | 
|  | SmallBitVector VFPRegsUsed; | 
|  | CfgVector<RegNumT> FP32Args; | 
|  | CfgVector<RegNumT> FP64Args; | 
|  | // UseFPRegs is a flag indicating if FP registers can be used | 
|  | bool UseFPRegs = false; | 
|  | }; | 
|  |  | 
|  | protected: | 
|  | explicit TargetMIPS32(Cfg *Func); | 
|  |  | 
|  | void postLower() override; | 
|  |  | 
|  | void lowerAlloca(const InstAlloca *Instr) override; | 
|  | void lowerArithmetic(const InstArithmetic *Instr) override; | 
|  | void lowerInt64Arithmetic(const InstArithmetic *Instr, Variable *Dest, | 
|  | Operand *Src0, Operand *Src1); | 
|  | void lowerAssign(const InstAssign *Instr) override; | 
|  | void lowerBr(const InstBr *Instr) override; | 
|  | void lowerBreakpoint(const InstBreakpoint *Instr) override; | 
|  | void lowerCall(const InstCall *Instr) override; | 
|  | void lowerCast(const InstCast *Instr) override; | 
|  | void lowerExtractElement(const InstExtractElement *Instr) override; | 
|  | void lowerFcmp(const InstFcmp *Instr) override; | 
|  | void lowerIcmp(const InstIcmp *Instr) override; | 
|  | void lower64Icmp(const InstIcmp *Instr); | 
|  | void createArithInst(Intrinsics::AtomicRMWOperation Operation, Variable *Dest, | 
|  | Variable *Src0, Variable *Src1); | 
|  | void lowerIntrinsic(const InstIntrinsic *Instr) override; | 
|  | void lowerInsertElement(const InstInsertElement *Instr) override; | 
|  | void lowerLoad(const InstLoad *Instr) override; | 
|  | void lowerPhi(const InstPhi *Instr) override; | 
|  | void lowerRet(const InstRet *Instr) override; | 
|  | void lowerSelect(const InstSelect *Instr) override; | 
|  | void lowerShuffleVector(const InstShuffleVector *Instr) override; | 
|  | void lowerStore(const InstStore *Instr) override; | 
|  | void lowerSwitch(const InstSwitch *Instr) override; | 
|  | void lowerUnreachable(const InstUnreachable *Instr) override; | 
|  | void lowerOther(const Inst *Instr) override; | 
|  | void prelowerPhis() override; | 
|  | uint32_t getCallStackArgumentsSizeBytes(const InstCall *Instr) override; | 
|  | void genTargetHelperCallFor(Inst *Instr) override; | 
|  | void doAddressOptLoad() override; | 
|  | void doAddressOptStore() override; | 
|  |  | 
|  | OperandMIPS32Mem *formMemoryOperand(Operand *Ptr, Type Ty); | 
|  |  | 
|  | class PostLoweringLegalizer { | 
|  | PostLoweringLegalizer() = delete; | 
|  | PostLoweringLegalizer(const PostLoweringLegalizer &) = delete; | 
|  | PostLoweringLegalizer &operator=(const PostLoweringLegalizer &) = delete; | 
|  |  | 
|  | public: | 
|  | explicit PostLoweringLegalizer(TargetMIPS32 *Target) | 
|  | : Target(Target), StackOrFrameReg(Target->getPhysicalRegister( | 
|  | Target->getFrameOrStackReg())) {} | 
|  |  | 
|  | /// Legalizes Mem. if Mem.Base is a rematerializable variable, | 
|  | /// Mem.Offset is fixed up. | 
|  | OperandMIPS32Mem *legalizeMemOperand(OperandMIPS32Mem *Mem); | 
|  |  | 
|  | /// Legalizes Immediate if larger value overflows range of 16 bits | 
|  | Variable *legalizeImmediate(int32_t Imm); | 
|  |  | 
|  | /// Legalizes Mov if its Source (or Destination) is a spilled Variable, or | 
|  | /// if its Source is a Rematerializable variable (this form is used in lieu | 
|  | /// of lea, which is not available in MIPS.) | 
|  | /// | 
|  | /// Moves to memory become store instructions, and moves from memory, loads. | 
|  | void legalizeMov(InstMIPS32Mov *Mov); | 
|  | void legalizeMovFp(InstMIPS32MovFP64ToI64 *MovInstr); | 
|  |  | 
|  | private: | 
|  | /// Creates a new Base register centered around [Base, +/- Offset]. | 
|  | Variable *newBaseRegister(Variable *Base, int32_t Offset, | 
|  | RegNumT ScratchRegNum); | 
|  |  | 
|  | TargetMIPS32 *const Target; | 
|  | Variable *const StackOrFrameReg; | 
|  | }; | 
|  |  | 
|  | bool UsesFramePointer = false; | 
|  | bool NeedsStackAlignment = false; | 
|  | bool MaybeLeafFunc = true; | 
|  | bool PrologEmitsFixedAllocas = false; | 
|  | bool VariableAllocaUsed = false; | 
|  | uint32_t MaxOutArgsSizeBytes = 0; | 
|  | uint32_t TotalStackSizeBytes = 0; | 
|  | uint32_t CurrentAllocaOffset = 0; | 
|  | uint32_t VariableAllocaAlignBytes = 0; | 
|  | static SmallBitVector TypeToRegisterSet[RCMIPS32_NUM]; | 
|  | static SmallBitVector TypeToRegisterSetUnfiltered[RCMIPS32_NUM]; | 
|  | static SmallBitVector RegisterAliases[RegMIPS32::Reg_NUM]; | 
|  | SmallBitVector RegsUsed; | 
|  | VarList PhysicalRegisters[IceType_NUM]; | 
|  | VarList PreservedGPRs; | 
|  | static constexpr uint32_t CHAR_BITS = 8; | 
|  | static constexpr uint32_t INT32_BITS = 32; | 
|  | size_t SpillAreaSizeBytes = 0; | 
|  | size_t FixedAllocaSizeBytes = 0; | 
|  | size_t FixedAllocaAlignBytes = 0; | 
|  | size_t PreservedRegsSizeBytes = 0; | 
|  | Variable *ImplicitRet = nullptr; /// Implicit return | 
|  |  | 
|  | private: | 
|  | ENABLE_MAKE_UNIQUE; | 
|  |  | 
|  | OperandMIPS32Mem *formAddressingMode(Type Ty, Cfg *Func, const Inst *LdSt, | 
|  | Operand *Base); | 
|  |  | 
|  | class ComputationTracker { | 
|  | public: | 
|  | ComputationTracker() = default; | 
|  | ~ComputationTracker() = default; | 
|  |  | 
|  | void forgetProducers() { KnownComputations.clear(); } | 
|  | void recordProducers(CfgNode *Node); | 
|  |  | 
|  | const Inst *getProducerOf(const Operand *Opnd) const { | 
|  | auto *Var = llvm::dyn_cast<Variable>(Opnd); | 
|  | if (Var == nullptr) { | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | auto Iter = KnownComputations.find(Var->getIndex()); | 
|  | if (Iter == KnownComputations.end()) { | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | return Iter->second.Instr; | 
|  | } | 
|  |  | 
|  | void dump(const Cfg *Func) const { | 
|  | if (!BuildDefs::dump() || !Func->isVerbose(IceV_Folding)) | 
|  | return; | 
|  | OstreamLocker L(Func->getContext()); | 
|  | Ostream &Str = Func->getContext()->getStrDump(); | 
|  | Str << "foldable producer:\n"; | 
|  | for (const auto &Computation : KnownComputations) { | 
|  | Str << "    "; | 
|  | Computation.second.Instr->dump(Func); | 
|  | Str << "\n"; | 
|  | } | 
|  | Str << "\n"; | 
|  | } | 
|  |  | 
|  | private: | 
|  | class ComputationEntry { | 
|  | public: | 
|  | ComputationEntry(Inst *I, Type Ty) : Instr(I), ComputationType(Ty) {} | 
|  | Inst *const Instr; | 
|  | // Boolean folding is disabled for variables whose live range is multi | 
|  | // block. We conservatively initialize IsLiveOut to true, and set it to | 
|  | // false once we find the end of the live range for the variable defined | 
|  | // by this instruction. If liveness analysis is not performed (e.g., in | 
|  | // Om1 mode) IsLiveOut will never be set to false, and folding will be | 
|  | // disabled. | 
|  | bool IsLiveOut = true; | 
|  | int32_t NumUses = 0; | 
|  | Type ComputationType; | 
|  | }; | 
|  |  | 
|  | // ComputationMap maps a Variable number to a payload identifying which | 
|  | // instruction defined it. | 
|  | using ComputationMap = CfgUnorderedMap<SizeT, ComputationEntry>; | 
|  | ComputationMap KnownComputations; | 
|  | }; | 
|  |  | 
|  | ComputationTracker Computations; | 
|  | }; | 
|  |  | 
|  | class TargetDataMIPS32 final : public TargetDataLowering { | 
|  | TargetDataMIPS32() = delete; | 
|  | TargetDataMIPS32(const TargetDataMIPS32 &) = delete; | 
|  | TargetDataMIPS32 &operator=(const TargetDataMIPS32 &) = delete; | 
|  |  | 
|  | public: | 
|  | static std::unique_ptr<TargetDataLowering> create(GlobalContext *Ctx) { | 
|  | return std::unique_ptr<TargetDataLowering>(new TargetDataMIPS32(Ctx)); | 
|  | } | 
|  |  | 
|  | void lowerGlobals(const VariableDeclarationList &Vars, | 
|  | const std::string &SectionSuffix) override; | 
|  | void lowerConstants() override; | 
|  | void lowerJumpTables() override; | 
|  | void emitTargetRODataSections() override; | 
|  |  | 
|  | protected: | 
|  | explicit TargetDataMIPS32(GlobalContext *Ctx); | 
|  |  | 
|  | private: | 
|  | ~TargetDataMIPS32() override = default; | 
|  | }; | 
|  |  | 
|  | class TargetHeaderMIPS32 final : public TargetHeaderLowering { | 
|  | TargetHeaderMIPS32() = delete; | 
|  | TargetHeaderMIPS32(const TargetHeaderMIPS32 &) = delete; | 
|  | TargetHeaderMIPS32 &operator=(const TargetHeaderMIPS32 &) = delete; | 
|  |  | 
|  | public: | 
|  | static std::unique_ptr<TargetHeaderLowering> create(GlobalContext *Ctx) { | 
|  | return std::unique_ptr<TargetHeaderLowering>(new TargetHeaderMIPS32(Ctx)); | 
|  | } | 
|  |  | 
|  | void lower() override; | 
|  |  | 
|  | protected: | 
|  | explicit TargetHeaderMIPS32(GlobalContext *Ctx); | 
|  |  | 
|  | private: | 
|  | ~TargetHeaderMIPS32() = default; | 
|  | }; | 
|  |  | 
|  | // This structure (with some minor modifications) is copied from | 
|  | // llvm/lib/Target/Mips/MCTargetDesc/MipsABIFlagsSection.h file. | 
|  | struct MipsABIFlagsSection { | 
|  |  | 
|  | // Version of the MIPS.abiflags section | 
|  | enum AFL_VERSION { | 
|  | AFL_VERSION_V0 = 0 // Version 0 | 
|  | }; | 
|  |  | 
|  | // The level of the ISA: 1-5, 32, 64. | 
|  | enum AFL_ISA_LEVEL { | 
|  | AFL_ISA_LEVEL_NONE = 0, | 
|  | AFL_ISA_LEVEL_MIPS32 = 32, // MIPS32 | 
|  | }; | 
|  |  | 
|  | // The revision of ISA: 0 for MIPS V and below, 1-n otherwise. | 
|  | enum AFL_ISA_REV { | 
|  | AFL_ISA_REV_NONE = 0, | 
|  | AFL_ISA_REV_R1 = 1, // R1 | 
|  | }; | 
|  |  | 
|  | // Values for the xxx_size bytes of an ABI flags structure. | 
|  | enum AFL_REG { | 
|  | AFL_REG_NONE = 0x00, // No registers. | 
|  | AFL_REG_32 = 0x01,   // 32-bit registers. | 
|  | AFL_REG_64 = 0x02,   // 64-bit registers. | 
|  | AFL_REG_128 = 0x03   // 128-bit registers. | 
|  | }; | 
|  |  | 
|  | // Values for the fp_abi word of an ABI flags structure. | 
|  | enum AFL_FP_ABI { | 
|  | AFL_FP_ANY = 0, | 
|  | AFL_FP_DOUBLE = 1, | 
|  | AFL_FP_XX = 5, | 
|  | AFL_FP_64 = 6, | 
|  | AFL_FP_64A = 7 | 
|  | }; | 
|  |  | 
|  | // Values for the isa_ext word of an ABI flags structure. | 
|  | enum AFL_EXT { | 
|  | AFL_EXT_NONE = 0, | 
|  | AFL_EXT_XLR = 1,          // RMI Xlr instruction. | 
|  | AFL_EXT_OCTEON2 = 2,      // Cavium Networks Octeon2. | 
|  | AFL_EXT_OCTEONP = 3,      // Cavium Networks OcteonP. | 
|  | AFL_EXT_LOONGSON_3A = 4,  // Loongson 3A. | 
|  | AFL_EXT_OCTEON = 5,       // Cavium Networks Octeon. | 
|  | AFL_EXT_5900 = 6,         // MIPS R5900 instruction. | 
|  | AFL_EXT_4650 = 7,         // MIPS R4650 instruction. | 
|  | AFL_EXT_4010 = 8,         // LSI R4010 instruction. | 
|  | AFL_EXT_4100 = 9,         // NEC VR4100 instruction. | 
|  | AFL_EXT_3900 = 10,        // Toshiba R3900 instruction. | 
|  | AFL_EXT_10000 = 11,       // MIPS R10000 instruction. | 
|  | AFL_EXT_SB1 = 12,         // Broadcom SB-1 instruction. | 
|  | AFL_EXT_4111 = 13,        // NEC VR4111/VR4181 instruction. | 
|  | AFL_EXT_4120 = 14,        // NEC VR4120 instruction. | 
|  | AFL_EXT_5400 = 15,        // NEC VR5400 instruction. | 
|  | AFL_EXT_5500 = 16,        // NEC VR5500 instruction. | 
|  | AFL_EXT_LOONGSON_2E = 17, // ST Microelectronics Loongson 2E. | 
|  | AFL_EXT_LOONGSON_2F = 18  // ST Microelectronics Loongson 2F. | 
|  | }; | 
|  |  | 
|  | // Masks for the ases word of an ABI flags structure. | 
|  | enum AFL_ASE { | 
|  | AFL_ASE_NONE = 0x00000000, | 
|  | AFL_ASE_DSP = 0x00000001,       // DSP ASE. | 
|  | AFL_ASE_DSPR2 = 0x00000002,     // DSP R2 ASE. | 
|  | AFL_ASE_EVA = 0x00000004,       // Enhanced VA Scheme. | 
|  | AFL_ASE_MCU = 0x00000008,       // MCU (MicroController) ASE. | 
|  | AFL_ASE_MDMX = 0x00000010,      // MDMX ASE. | 
|  | AFL_ASE_MIPS3D = 0x00000020,    // MIPS-3D ASE. | 
|  | AFL_ASE_MT = 0x00000040,        // MT ASE. | 
|  | AFL_ASE_SMARTMIPS = 0x00000080, // SmartMIPS ASE. | 
|  | AFL_ASE_VIRT = 0x00000100,      // VZ ASE. | 
|  | AFL_ASE_MSA = 0x00000200,       // MSA ASE. | 
|  | AFL_ASE_MIPS16 = 0x00000400,    // MIPS16 ASE. | 
|  | AFL_ASE_MICROMIPS = 0x00000800, // MICROMIPS ASE. | 
|  | AFL_ASE_XPA = 0x00001000        // XPA ASE. | 
|  | }; | 
|  |  | 
|  | enum AFL_FLAGS1 { AFL_FLAGS1_NONE = 0, AFL_FLAGS1_ODDSPREG = 1 }; | 
|  |  | 
|  | enum AFL_FLAGS2 { AFL_FLAGS2_NONE = 0 }; | 
|  |  | 
|  | uint16_t Version = AFL_VERSION_V0; | 
|  | uint8_t ISALevel = AFL_ISA_LEVEL_MIPS32; | 
|  | uint8_t ISARevision = AFL_ISA_REV_R1; | 
|  | uint8_t GPRSize = AFL_REG_32; | 
|  | uint8_t CPR1Size = AFL_REG_32; | 
|  | uint8_t CPR2Size = AFL_REG_NONE; | 
|  | uint8_t FPABI = AFL_FP_DOUBLE; | 
|  | uint32_t Extension = AFL_EXT_NONE; | 
|  | uint32_t ASE = AFL_ASE_NONE; | 
|  | uint32_t Flags1 = AFL_FLAGS1_ODDSPREG; | 
|  | uint32_t Flags2 = AFL_FLAGS2_NONE; | 
|  |  | 
|  | MipsABIFlagsSection() = default; | 
|  | }; | 
|  |  | 
|  | } // end of namespace MIPS32 | 
|  | } // end of namespace Ice | 
|  |  | 
|  | #endif // SUBZERO_SRC_ICETARGETLOWERINGMIPS32_H |