blob: c93c0615158ca1eae2f0a43d8fc115bc1413806f [file] [log] [blame]
//===- subzero/unittest/unittest/AssemblerX8664/TestUtil.h ------*- C++ -*-===//
//
// The Subzero Code Generator
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
//
// Utility classes for testing the X8664 Assembler.
//
//===----------------------------------------------------------------------===//
#ifndef ASSEMBLERX8664_TESTUTIL_H_
#define ASSEMBLERX8664_TESTUTIL_H_
#include "IceAssemblerX8664.h"
#include "gtest/gtest.h"
#if defined(__unix__)
#include <sys/mman.h>
#elif defined(_WIN32)
#define NOMINMAX
#include <Windows.h>
#else
#error "Platform unsupported"
#endif
#include <cassert>
namespace Ice {
namespace X8664 {
namespace Test {
class AssemblerX8664TestBase : public ::testing::Test {
protected:
using Address = AssemblerX8664::Traits::Address;
using Cond = AssemblerX8664::CondX86;
using GPRRegister = AssemblerX8664::Traits::GPRRegister;
using ByteRegister = AssemblerX8664::Traits::ByteRegister;
using Traits = AssemblerX8664::Traits;
using XmmRegister = AssemblerX8664::Traits::XmmRegister;
// The following are "nicknames" for all possible GPRs in x86-64. With those, we
// can use, e.g.,
//
// Encoded_GPR_al()
//
// instead of GPRRegister::Encoded_Reg_eax for 8 bit operands. They also
// introduce "regular" nicknames for legacy x86-32 register (e.g., eax becomes
// r1; esp, r0).
#define LegacyRegAliases(NewName, Name64, Name32, Name16, Name8) \
static constexpr GPRRegister Encoded_GPR_##NewName() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##NewName##q() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##NewName##d() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##NewName##w() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##NewName##l() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr ByteRegister Encoded_Bytereg_##NewName() { \
return ByteRegister::Encoded_8_Reg_##Name8; \
} \
static constexpr GPRRegister Encoded_GPR_##Name64() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##Name32() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##Name16() { \
return GPRRegister::Encoded_Reg_##Name32; \
} \
static constexpr GPRRegister Encoded_GPR_##Name8() { \
return GPRRegister::Encoded_Reg_##Name32; \
}
#define NewRegAliases(Name) \
static constexpr GPRRegister Encoded_GPR_##Name() { \
return GPRRegister::Encoded_Reg_##Name##d; \
} \
static constexpr GPRRegister Encoded_GPR_##Name##q() { \
return GPRRegister::Encoded_Reg_##Name##d; \
} \
static constexpr GPRRegister Encoded_GPR_##Name##d() { \
return GPRRegister::Encoded_Reg_##Name##d; \
} \
static constexpr GPRRegister Encoded_GPR_##Name##w() { \
return GPRRegister::Encoded_Reg_##Name##d; \
} \
static constexpr GPRRegister Encoded_GPR_##Name##l() { \
return GPRRegister::Encoded_Reg_##Name##d; \
} \
static constexpr ByteRegister Encoded_Bytereg_##Name() { \
return ByteRegister::Encoded_8_Reg_##Name##l; \
}
#define XmmRegAliases(Name) \
static constexpr XmmRegister Encoded_Xmm_##Name() { \
return XmmRegister::Encoded_Reg_##Name; \
}
LegacyRegAliases(r0, rsp, esp, sp, spl);
LegacyRegAliases(r1, rax, eax, ax, al);
LegacyRegAliases(r2, rbx, ebx, bx, bl);
LegacyRegAliases(r3, rcx, ecx, cx, cl);
LegacyRegAliases(r4, rdx, edx, dx, dl);
LegacyRegAliases(r5, rbp, ebp, bp, bpl);
LegacyRegAliases(r6, rsi, esi, si, sil);
LegacyRegAliases(r7, rdi, edi, di, dil);
NewRegAliases(r8);
NewRegAliases(r9);
NewRegAliases(r10);
NewRegAliases(r11);
NewRegAliases(r12);
NewRegAliases(r13);
NewRegAliases(r14);
NewRegAliases(r15);
XmmRegAliases(xmm0);
XmmRegAliases(xmm1);
XmmRegAliases(xmm2);
XmmRegAliases(xmm3);
XmmRegAliases(xmm4);
XmmRegAliases(xmm5);
XmmRegAliases(xmm6);
XmmRegAliases(xmm7);
XmmRegAliases(xmm8);
XmmRegAliases(xmm9);
XmmRegAliases(xmm10);
XmmRegAliases(xmm11);
XmmRegAliases(xmm12);
XmmRegAliases(xmm13);
XmmRegAliases(xmm14);
XmmRegAliases(xmm15);
#undef XmmRegAliases
#undef NewRegAliases
#undef LegacyRegAliases
AssemblerX8664TestBase() { reset(); }
void reset() { Assembler = makeUnique<AssemblerX8664>(); }
AssemblerX8664 *assembler() const { return Assembler.get(); }
size_t codeBytesSize() const { return Assembler->getBufferView().size(); }
const uint8_t *codeBytes() const {
return static_cast<const uint8_t *>(
static_cast<const void *>(Assembler->getBufferView().data()));
}
private:
std::unique_ptr<AssemblerX8664> Assembler;
};
// __ is a helper macro. It allows test cases to emit X8664 assembly
// instructions with
//
// __ mov(GPRRegister::Reg_Eax, 1);
// __ ret();
//
// and so on. The idea of having this was "stolen" from dart's unit tests.
#define __ (this->assembler())->
// AssemblerX8664LowLevelTest verify that the "basic" instructions the tests
// rely on are encoded correctly. Therefore, instead of executing the assembled
// code, these tests will verify that the assembled bytes are sane.
class AssemblerX8664LowLevelTest : public AssemblerX8664TestBase {
protected:
// verifyBytes is a template helper that takes a Buffer, and a variable number
// of bytes. As the name indicates, it is used to verify the bytes for an
// instruction encoding.
template <int N, int I> static bool verifyBytes(const uint8_t *) {
static_assert(I == N, "Invalid template instantiation.");
return true;
}
template <int N, int I = 0, typename... Args>
static bool verifyBytes(const uint8_t *Buffer, uint8_t Byte,
Args... OtherBytes) {
static_assert(I < N, "Invalid template instantiation.");
EXPECT_EQ(Byte, Buffer[I]) << "Byte " << (I + 1) << " of " << N;
return verifyBytes<N, I + 1>(Buffer, OtherBytes...) && Buffer[I] == Byte;
}
};
// After these tests we should have a sane environment; we know the following
// work:
//
// (*) zeroing eax, ebx, ecx, edx, edi, and esi;
// (*) call $4 instruction (used for ip materialization);
// (*) register push and pop;
// (*) cmp reg, reg; and
// (*) returning from functions.
//
// We can now dive into testing each emitting method in AssemblerX8664. Each
// test will emit some instructions for performing the test. The assembled
// instructions will operate in a "safe" environment. All x86-64 registers are
// spilled to the program stack, and the registers are then zeroed out, with the
// exception of %esp and %r9.
//
// The jitted code and the unittest code will share the same stack. Therefore,
// test harnesses need to ensure it does not leave anything it pushed on the
// stack.
//
// %r9 is initialized with a pointer for rIP-based addressing. This pointer is
// used for position-independent access to a scratchpad area for use in tests.
// In theory we could use rip-based addressing, but in practice that would
// require creating fixups, which would, in turn, require creating a global
// context. We therefore rely on the same technique used for pic code in x86-32
// (i.e., IP materialization). Upon a test start up, a call(NextInstruction) is
// executed. We then pop the return address from the stack, and use it for pic
// addressing.
//
// The jitted code will look like the following:
//
// test:
// push %r9
// call test$materialize_ip
// test$materialize_ip: <<------- %r9 will point here
// pop %r9
// push %rax
// push %rbx
// push %rcx
// push %rdx
// push %rbp
// push %rdi
// push %rsi
// push %r8
// push %r10
// push %r11
// push %r12
// push %r13
// push %r14
// push %r15
// mov $0, %rax
// mov $0, %rbx
// mov $0, %rcx
// mov $0, %rdx
// mov $0, %rbp
// mov $0, %rdi
// mov $0, %rsi
// mov $0, %r8
// mov $0, %r10
// mov $0, %r11
// mov $0, %r12
// mov $0, %r13
// mov $0, %r14
// mov $0, %r15
//
// << test code goes here >>
//
// mov %rax, { 0 + $ScratchpadOffset}(%rbp)
// mov %rbx, { 8 + $ScratchpadOffset}(%rbp)
// mov %rcx, { 16 + $ScratchpadOffset}(%rbp)
// mov %rdx, { 24 + $ScratchpadOffset}(%rbp)
// mov %rdi, { 32 + $ScratchpadOffset}(%rbp)
// mov %rsi, { 40 + $ScratchpadOffset}(%rbp)
// mov %rbp, { 48 + $ScratchpadOffset}(%rbp)
// mov %rsp, { 56 + $ScratchpadOffset}(%rbp)
// mov %r8, { 64 + $ScratchpadOffset}(%rbp)
// mov %r9, { 72 + $ScratchpadOffset}(%rbp)
// mov %r10, { 80 + $ScratchpadOffset}(%rbp)
// mov %r11, { 88 + $ScratchpadOffset}(%rbp)
// mov %r12, { 96 + $ScratchpadOffset}(%rbp)
// mov %r13, {104 + $ScratchpadOffset}(%rbp)
// mov %r14, {112 + $ScratchpadOffset}(%rbp)
// mov %r15, {120 + $ScratchpadOffset}(%rbp)
// movups %xmm0, {128 + $ScratchpadOffset}(%rbp)
// movups %xmm1, {136 + $ScratchpadOffset}(%rbp)
// movups %xmm2, {144 + $ScratchpadOffset}(%rbp)
// movups %xmm3, {152 + $ScratchpadOffset}(%rbp)
// movups %xmm4, {160 + $ScratchpadOffset}(%rbp)
// movups %xmm5, {168 + $ScratchpadOffset}(%rbp)
// movups %xmm6, {176 + $ScratchpadOffset}(%rbp)
// movups %xmm7, {184 + $ScratchpadOffset}(%rbp)
// movups %xmm8, {192 + $ScratchpadOffset}(%rbp)
// movups %xmm9, {200 + $ScratchpadOffset}(%rbp)
// movups %xmm10, {208 + $ScratchpadOffset}(%rbp)
// movups %xmm11, {216 + $ScratchpadOffset}(%rbp)
// movups %xmm12, {224 + $ScratchpadOffset}(%rbp)
// movups %xmm13, {232 + $ScratchpadOffset}(%rbp)
// movups %xmm14, {240 + $ScratchpadOffset}(%rbp)
// movups %xmm15, {248 + $ScratchpadOffset}(%rbp)
//
// pop %r15
// pop %r14
// pop %r13
// pop %r12
// pop %r11
// pop %r10
// pop %r8
// pop %rsi
// pop %rdi
// pop %rbp
// pop %rdx
// pop %rcx
// pop %rbx
// pop %rax
// pop %r9
// ret
//
// << ... >>
//
// scratchpad: <<------- accessed via $Offset(%ebp)
//
// << test scratch area >>
//
// TODO(jpp): test the
//
// mov %reg, $Offset(%ebp)
// movups %xmm, $Offset(%ebp)
//
// encodings using the low level assembler test ensuring that the register
// values can be written to the scratchpad area.
//
// r9 was deliberately choosen so that every instruction accessing memory would
// fail if the rex prefix was not emitted for it.
class AssemblerX8664Test : public AssemblerX8664TestBase {
protected:
// Dqword is used to represent 128-bit data types. The Dqword's contents are
// the same as the contents read from memory. Tests can then use the union
// members to verify the tests' outputs.
//
// NOTE: We want sizeof(Dqword) == sizeof(uint64_t) * 2. In other words, we
// want Dqword's contents to be **exactly** what the memory contents were so
// that we can do, e.g.,
//
// ...
// float Ret[4];
// // populate Ret
// return *reinterpret_cast<Dqword *>(&Ret);
//
// While being an ugly hack, this kind of return statements are used
// extensively in the PackedArith (see below) class.
union Dqword {
template <typename T0, typename T1, typename T2, typename T3,
typename = typename std::enable_if<
std::is_floating_point<T0>::value>::type>
Dqword(T0 F0, T1 F1, T2 F2, T3 F3) {
F32[0] = F0;
F32[1] = F1;
F32[2] = F2;
F32[3] = F3;
}
template <typename T>
Dqword(typename std::enable_if<std::is_same<T, int32_t>::value, T>::type I0,
T I1, T I2, T I3) {
I32[0] = I0;
I32[1] = I1;
I32[2] = I2;
I32[3] = I3;
}
template <typename T>
Dqword(typename std::enable_if<std::is_same<T, uint64_t>::value, T>::type
U64_0,
T U64_1) {
U64[0] = U64_0;
U64[1] = U64_1;
}
template <typename T>
Dqword(typename std::enable_if<std::is_same<T, double>::value, T>::type D0,
T D1) {
F64[0] = D0;
F64[1] = D1;
}
bool operator==(const Dqword &Rhs) const {
return std::memcmp(this, &Rhs, sizeof(*this)) == 0;
}
double F64[2];
uint64_t U64[2];
int64_t I64[2];
float F32[4];
uint32_t U32[4];
int32_t I32[4];
uint16_t U16[8];
int16_t I16[8];
uint8_t U8[16];
int8_t I8[16];
private:
Dqword() = delete;
};
// As stated, we want this condition to hold, so we assert.
static_assert(sizeof(Dqword) == 2 * sizeof(uint64_t),
"Dqword has the wrong size.");
// PackedArith is an interface provider for Dqwords. PackedArith's C argument
// is the undelying Dqword's type, which is then used so that we can define
// operators in terms of C++ operators on the underlying elements' type.
template <typename C> class PackedArith {
public:
static constexpr uint32_t N = sizeof(Dqword) / sizeof(C);
static_assert(N * sizeof(C) == sizeof(Dqword),
"Invalid template paramenter.");
static_assert((N & 1) == 0, "N should be divisible by 2");
#define DefinePackedComparisonOperator(Op) \
template <typename Container = C, int Size = N> \
typename std::enable_if<std::is_floating_point<Container>::value, \
Dqword>::type \
operator Op(const Dqword &Rhs) const { \
using ElemType = \
typename std::conditional<std::is_same<float, Container>::value, \
int32_t, int64_t>::type; \
static_assert(sizeof(ElemType) == sizeof(Container), \
"Check ElemType definition."); \
const ElemType *const RhsPtr = \
reinterpret_cast<const ElemType *const>(&Rhs); \
const ElemType *const LhsPtr = \
reinterpret_cast<const ElemType *const>(&Lhs); \
ElemType Ret[N]; \
for (uint32_t i = 0; i < N; ++i) { \
Ret[i] = (LhsPtr[i] Op RhsPtr[i]) ? -1 : 0; \
} \
return *reinterpret_cast<Dqword *>(&Ret); \
}
DefinePackedComparisonOperator(<);
DefinePackedComparisonOperator(<=);
DefinePackedComparisonOperator(>);
DefinePackedComparisonOperator(>=);
DefinePackedComparisonOperator(==);
DefinePackedComparisonOperator(!=);
#undef DefinePackedComparisonOperator
#define DefinePackedOrdUnordComparisonOperator(Op, Ordered) \
template <typename Container = C, int Size = N> \
typename std::enable_if<std::is_floating_point<Container>::value, \
Dqword>::type \
Op(const Dqword &Rhs) const { \
using ElemType = \
typename std::conditional<std::is_same<float, Container>::value, \
int32_t, int64_t>::type; \
static_assert(sizeof(ElemType) == sizeof(Container), \
"Check ElemType definition."); \
const Container *const RhsPtr = \
reinterpret_cast<const Container *const>(&Rhs); \
const Container *const LhsPtr = \
reinterpret_cast<const Container *const>(&Lhs); \
ElemType Ret[N]; \
for (uint32_t i = 0; i < N; ++i) { \
Ret[i] = (!(LhsPtr[i] == LhsPtr[i]) || !(RhsPtr[i] == RhsPtr[i])) != \
(Ordered) \
? -1 \
: 0; \
} \
return *reinterpret_cast<Dqword *>(&Ret); \
}
DefinePackedOrdUnordComparisonOperator(ord, true);
DefinePackedOrdUnordComparisonOperator(unord, false);
#undef DefinePackedOrdUnordComparisonOperator
#define DefinePackedArithOperator(Op, RhsIndexChanges, NeedsInt) \
template <typename Container = C, int Size = N> \
Dqword operator Op(const Dqword &Rhs) const { \
using ElemTypeForFp = typename std::conditional< \
!(NeedsInt), Container, \
typename std::conditional< \
std::is_same<Container, float>::value, uint32_t, \
typename std::conditional<std::is_same<Container, double>::value, \
uint64_t, void>::type>::type>::type; \
using ElemType = \
typename std::conditional<std::is_integral<Container>::value, \
Container, ElemTypeForFp>::type; \
static_assert(!std::is_same<void, ElemType>::value, \
"Check ElemType definition."); \
const ElemType *const RhsPtr = \
reinterpret_cast<const ElemType *const>(&Rhs); \
const ElemType *const LhsPtr = \
reinterpret_cast<const ElemType *const>(&Lhs); \
ElemType Ret[N]; \
for (uint32_t i = 0; i < N; ++i) { \
Ret[i] = LhsPtr[i] Op RhsPtr[(RhsIndexChanges) ? i : 0]; \
} \
return *reinterpret_cast<Dqword *>(&Ret); \
}
DefinePackedArithOperator(>>, false, true);
DefinePackedArithOperator(<<, false, true);
DefinePackedArithOperator(+, true, false);
DefinePackedArithOperator(-, true, false);
DefinePackedArithOperator(/, true, false);
DefinePackedArithOperator(&, true, true);
DefinePackedArithOperator(|, true, true);
DefinePackedArithOperator(^, true, true);
#undef DefinePackedArithOperator
#define DefinePackedArithShiftImm(Op) \
template <typename Container = C, int Size = N> \
Dqword operator Op(uint8_t imm) const { \
const Container *const LhsPtr = \
reinterpret_cast<const Container *const>(&Lhs); \
Container Ret[N]; \
for (uint32_t i = 0; i < N; ++i) { \
Ret[i] = LhsPtr[i] Op imm; \
} \
return *reinterpret_cast<Dqword *>(&Ret); \
}
DefinePackedArithShiftImm(>>);
DefinePackedArithShiftImm(<<);
#undef DefinePackedArithShiftImm
template <typename Container = C, int Size = N>
typename std::enable_if<std::is_signed<Container>::value ||
std::is_floating_point<Container>::value,
Dqword>::type
operator*(const Dqword &Rhs) const {
static_assert((std::is_integral<Container>::value &&
sizeof(Container) < sizeof(uint64_t)) ||
std::is_floating_point<Container>::value,
"* is only defined for i(8|16|32), and fp types.");
const Container *const RhsPtr =
reinterpret_cast<const Container *const>(&Rhs);
const Container *const LhsPtr =
reinterpret_cast<const Container *const>(&Lhs);
Container Ret[Size];
for (uint32_t i = 0; i < Size; ++i) {
Ret[i] = LhsPtr[i] * RhsPtr[i];
}
return *reinterpret_cast<Dqword *>(&Ret);
}
template <typename Container = C, int Size = N,
typename = typename std::enable_if<
!std::is_signed<Container>::value>::type>
Dqword operator*(const Dqword &Rhs) const {
static_assert(std::is_integral<Container>::value &&
sizeof(Container) < sizeof(uint64_t),
"* is only defined for ui(8|16|32)");
using NextType = typename std::conditional<
sizeof(Container) == 1, uint16_t,
typename std::conditional<sizeof(Container) == 2, uint32_t,
uint64_t>::type>::type;
static_assert(sizeof(Container) * 2 == sizeof(NextType),
"Unexpected size");
const Container *const RhsPtr =
reinterpret_cast<const Container *const>(&Rhs);
const Container *const LhsPtr =
reinterpret_cast<const Container *const>(&Lhs);
NextType Ret[Size / 2];
for (uint32_t i = 0; i < Size; i += 2) {
Ret[i / 2] =
static_cast<NextType>(LhsPtr[i]) * static_cast<NextType>(RhsPtr[i]);
}
return *reinterpret_cast<Dqword *>(&Ret);
}
template <typename Container = C, int Size = N>
PackedArith<Container> operator~() const {
const Container *const LhsPtr =
reinterpret_cast<const Container *const>(&Lhs);
Container Ret[Size];
for (uint32_t i = 0; i < Size; ++i) {
Ret[i] = ~LhsPtr[i];
}
return PackedArith<Container>(*reinterpret_cast<Dqword *>(&Ret));
}
#define MinMaxOperations(Name, Suffix) \
template <typename Container = C, int Size = N> \
Dqword Name##Suffix(const Dqword &Rhs) const { \
static_assert(std::is_floating_point<Container>::value, \
#Name #Suffix "ps is only available for fp."); \
const Container *const RhsPtr = \
reinterpret_cast<const Container *const>(&Rhs); \
const Container *const LhsPtr = \
reinterpret_cast<const Container *const>(&Lhs); \
Container Ret[Size]; \
for (uint32_t i = 0; i < Size; ++i) { \
Ret[i] = std::Name(LhsPtr[i], RhsPtr[i]); \
} \
return *reinterpret_cast<Dqword *>(&Ret); \
}
MinMaxOperations(max, ps);
MinMaxOperations(max, pd);
MinMaxOperations(min, ps);
MinMaxOperations(min, pd);
#undef MinMaxOperations
template <typename Container = C, int Size = N>
Dqword blendWith(const Dqword &Rhs, const Dqword &Mask) const {
using MaskType = typename std::conditional<
sizeof(Container) == 1, int8_t,
typename std::conditional<sizeof(Container) == 2, int16_t,
int32_t>::type>::type;
static_assert(sizeof(MaskType) == sizeof(Container),
"MaskType has the wrong size.");
const Container *const RhsPtr =
reinterpret_cast<const Container *const>(&Rhs);
const Container *const LhsPtr =
reinterpret_cast<const Container *const>(&Lhs);
const MaskType *const MaskPtr =
reinterpret_cast<const MaskType *const>(&Mask);
Container Ret[Size];
for (int i = 0; i < Size; ++i) {
Ret[i] = ((MaskPtr[i] < 0) ? RhsPtr : LhsPtr)[i];
}
return *reinterpret_cast<Dqword *>(&Ret);
}
private:
// The AssemblerX8664Test class needs to be a friend so that it can create
// PackedArith objects (see below.)
friend class AssemblerX8664Test;
explicit PackedArith(const Dqword &MyLhs) : Lhs(MyLhs) {}
// Lhs can't be a & because operator~ returns a temporary object that needs
// access to its own Dqword.
const Dqword Lhs;
};
// Named constructor for PackedArith objects.
template <typename C> static PackedArith<C> packedAs(const Dqword &D) {
return PackedArith<C>(D);
}
AssemblerX8664Test() { reset(); }
void reset() {
AssemblerX8664TestBase::reset();
NeedsEpilogue = true;
// These dwords are allocated for saving the GPR state after the jitted code
// runs.
NumAllocatedDwords = AssembledTest::ScratchpadSlots;
addPrologue();
}
// AssembledTest is a wrapper around a PROT_EXEC mmap'ed buffer. This buffer
// contains both the test code as well as prologue/epilogue, and the
// scratchpad area that tests may use -- all tests use this scratchpad area
// for storing the processor's registers after the tests executed. This class
// also exposes helper methods for reading the register state after test
// execution, as well as for reading the scratchpad area.
class AssembledTest {
AssembledTest() = delete;
AssembledTest(const AssembledTest &) = delete;
AssembledTest &operator=(const AssembledTest &) = delete;
public:
static constexpr uint32_t MaximumCodeSize = 1 << 20;
static constexpr uint32_t raxSlot() { return 0; }
static constexpr uint32_t rbxSlot() { return 2; }
static constexpr uint32_t rcxSlot() { return 4; }
static constexpr uint32_t rdxSlot() { return 6; }
static constexpr uint32_t rdiSlot() { return 8; }
static constexpr uint32_t rsiSlot() { return 10; }
static constexpr uint32_t rbpSlot() { return 12; }
static constexpr uint32_t rspSlot() { return 14; }
static constexpr uint32_t r8Slot() { return 16; }
static constexpr uint32_t r9Slot() { return 18; }
static constexpr uint32_t r10Slot() { return 20; }
static constexpr uint32_t r11Slot() { return 22; }
static constexpr uint32_t r12Slot() { return 24; }
static constexpr uint32_t r13Slot() { return 26; }
static constexpr uint32_t r14Slot() { return 28; }
static constexpr uint32_t r15Slot() { return 30; }
// save 4 dwords for each xmm registers.
static constexpr uint32_t xmm0Slot() { return 32; }
static constexpr uint32_t xmm1Slot() { return 36; }
static constexpr uint32_t xmm2Slot() { return 40; }
static constexpr uint32_t xmm3Slot() { return 44; }
static constexpr uint32_t xmm4Slot() { return 48; }
static constexpr uint32_t xmm5Slot() { return 52; }
static constexpr uint32_t xmm6Slot() { return 56; }
static constexpr uint32_t xmm7Slot() { return 60; }
static constexpr uint32_t xmm8Slot() { return 64; }
static constexpr uint32_t xmm9Slot() { return 68; }
static constexpr uint32_t xmm10Slot() { return 72; }
static constexpr uint32_t xmm11Slot() { return 76; }
static constexpr uint32_t xmm12Slot() { return 80; }
static constexpr uint32_t xmm13Slot() { return 84; }
static constexpr uint32_t xmm14Slot() { return 88; }
static constexpr uint32_t xmm15Slot() { return 92; }
static constexpr uint32_t ScratchpadSlots = 96;
AssembledTest(const uint8_t *Data, const size_t MySize,
const size_t ExtraStorageDwords)
: Size(MaximumCodeSize + 4 * ExtraStorageDwords) {
// MaxCodeSize is needed because EXPECT_LT needs a symbol with a name --
// probably a compiler bug?
uint32_t MaxCodeSize = MaximumCodeSize;
EXPECT_LT(MySize, MaxCodeSize);
assert(MySize < MaximumCodeSize);
#if defined(__unix__)
ExecutableData = mmap(nullptr, Size, PROT_WRITE | PROT_READ | PROT_EXEC,
MAP_PRIVATE | MAP_ANONYMOUS | MAP_32BIT, -1, 0);
EXPECT_NE(MAP_FAILED, ExecutableData) << strerror(errno);
assert(MAP_FAILED != ExecutableData);
#elif defined(_WIN32)
ExecutableData = VirtualAlloc(NULL, Size, MEM_COMMIT | MEM_RESERVE,
PAGE_EXECUTE_READWRITE);
EXPECT_NE(nullptr, ExecutableData) << strerror(errno);
assert(nullptr != ExecutableData);
#else
#error "Platform unsupported"
#endif
std::memcpy(ExecutableData, Data, MySize);
}
// We allow AssembledTest to be moved so that we can return objects of
// this type.
AssembledTest(AssembledTest &&Buffer)
: ExecutableData(Buffer.ExecutableData), Size(Buffer.Size) {
Buffer.ExecutableData = nullptr;
Buffer.Size = 0;
}
AssembledTest &operator=(AssembledTest &&Buffer) {
ExecutableData = Buffer.ExecutableData;
Buffer.ExecutableData = nullptr;
Size = Buffer.Size;
Buffer.Size = 0;
return *this;
}
~AssembledTest() {
if (ExecutableData != nullptr) {
#if defined(__unix__)
munmap(ExecutableData, Size);
#elif defined(_WIN32)
VirtualFree(ExecutableData, 0, MEM_RELEASE);
#else
#error "Platform unsupported"
#endif
ExecutableData = nullptr;
}
}
void run() const { reinterpret_cast<void (*)()>(ExecutableData)(); }
#define LegacyRegAccessors(NewName, Name64, Name32, Name16, Name8) \
static_assert(Encoded_GPR_##NewName() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint64_t NewName() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##NewName##q() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint64_t NewName##q() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##NewName##d() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint32_t NewName##d() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##NewName##w() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint16_t NewName##w() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##NewName##l() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint8_t NewName##l() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##Name64() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint64_t Name64() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##Name32() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint32_t Name32() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##Name16() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint16_t Name16() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
} \
static_assert(Encoded_GPR_##Name8() == Encoded_GPR_##Name64(), \
"Invalid aliasing."); \
uint8_t Name8() const { \
return contentsOfQword(AssembledTest::Name64##Slot()); \
}
#define NewRegAccessors(NewName) \
uint64_t NewName() const { \
return contentsOfQword(AssembledTest::NewName##Slot()); \
} \
uint64_t NewName##q() const { \
return contentsOfQword(AssembledTest::NewName##Slot()); \
} \
uint32_t NewName##d() const { \
return contentsOfQword(AssembledTest::NewName##Slot()); \
} \
uint16_t NewName##w() const { \
return contentsOfQword(AssembledTest::NewName##Slot()); \
} \
uint8_t NewName##l() const { \
return contentsOfQword(AssembledTest::NewName##Slot()); \
}
#define XmmRegAccessor(Name) \
template <typename T> T Name() const { \
return xmm<T>(AssembledTest::Name##Slot()); \
}
LegacyRegAccessors(r0, rsp, esp, sp, spl);
LegacyRegAccessors(r1, rax, eax, ax, al);
LegacyRegAccessors(r2, rbx, ebx, bx, bl);
LegacyRegAccessors(r3, rcx, ecx, cx, cl);
LegacyRegAccessors(r4, rdx, edx, dx, dl);
LegacyRegAccessors(r5, rbp, ebp, bp, bpl);
LegacyRegAccessors(r6, rsi, esi, si, sil);
LegacyRegAccessors(r7, rdi, edi, di, dil);
NewRegAccessors(r8);
NewRegAccessors(r9);
NewRegAccessors(r10);
NewRegAccessors(r11);
NewRegAccessors(r12);
NewRegAccessors(r13);
NewRegAccessors(r14);
NewRegAccessors(r15);
XmmRegAccessor(xmm0);
XmmRegAccessor(xmm1);
XmmRegAccessor(xmm2);
XmmRegAccessor(xmm3);
XmmRegAccessor(xmm4);
XmmRegAccessor(xmm5);
XmmRegAccessor(xmm6);
XmmRegAccessor(xmm7);
XmmRegAccessor(xmm8);
XmmRegAccessor(xmm9);
XmmRegAccessor(xmm10);
XmmRegAccessor(xmm11);
XmmRegAccessor(xmm12);
XmmRegAccessor(xmm13);
XmmRegAccessor(xmm14);
XmmRegAccessor(xmm15);
#undef XmmRegAccessor
#undef NewRegAccessors
#undef LegacyRegAccessors
// contentsOfDword is used for reading the values in the scratchpad area.
// Valid arguments are the dword ids returned by
// AssemblerX8664Test::allocateDword() -- other inputs are considered
// invalid, and are not guaranteed to work if the implementation changes.
template <typename T = uint32_t, typename = typename std::enable_if<
sizeof(T) == sizeof(uint32_t)>::type>
T contentsOfDword(uint32_t Dword) const {
return *reinterpret_cast<T *>(static_cast<uint8_t *>(ExecutableData) +
dwordOffset(Dword));
}
template <typename T = uint64_t, typename = typename std::enable_if<
sizeof(T) == sizeof(uint64_t)>::type>
T contentsOfQword(uint32_t InitialDword) const {
return *reinterpret_cast<T *>(static_cast<uint8_t *>(ExecutableData) +
dwordOffset(InitialDword));
}
Dqword contentsOfDqword(uint32_t InitialDword) const {
return *reinterpret_cast<Dqword *>(
static_cast<uint8_t *>(ExecutableData) + dwordOffset(InitialDword));
}
template <typename T = uint32_t, typename = typename std::enable_if<
sizeof(T) == sizeof(uint32_t)>::type>
void setDwordTo(uint32_t Dword, T value) {
*reinterpret_cast<uint32_t *>(static_cast<uint8_t *>(ExecutableData) +
dwordOffset(Dword)) =
*reinterpret_cast<uint32_t *>(&value);
}
template <typename T = uint64_t, typename = typename std::enable_if<
sizeof(T) == sizeof(uint64_t)>::type>
void setQwordTo(uint32_t InitialDword, T value) {
*reinterpret_cast<uint64_t *>(static_cast<uint8_t *>(ExecutableData) +
dwordOffset(InitialDword)) =
*reinterpret_cast<uint64_t *>(&value);
}
void setDqwordTo(uint32_t InitialDword, const Dqword &qdword) {
setQwordTo(InitialDword, qdword.U64[0]);
setQwordTo(InitialDword + 2, qdword.U64[1]);
}
private:
template <typename T>
typename std::enable_if<std::is_same<T, Dqword>::value, Dqword>::type
xmm(uint8_t Slot) const {
return contentsOfDqword(Slot);
}
template <typename T>
typename std::enable_if<!std::is_same<T, Dqword>::value, T>::type
xmm(uint8_t Slot) const {
constexpr bool TIs64Bit = sizeof(T) == sizeof(uint64_t);
using _64BitType = typename std::conditional<TIs64Bit, T, uint64_t>::type;
using _32BitType = typename std::conditional<TIs64Bit, uint32_t, T>::type;
if (TIs64Bit) {
return contentsOfQword<_64BitType>(Slot);
}
return contentsOfDword<_32BitType>(Slot);
}
static uint32_t dwordOffset(uint32_t Index) {
return MaximumCodeSize + (Index * 4);
}
void *ExecutableData = nullptr;
size_t Size;
};
// assemble created an AssembledTest with the jitted code. The first time
// assemble is executed it will add the epilogue to the jitted code (which is
// the reason why this method is not const qualified.
AssembledTest assemble() {
if (NeedsEpilogue) {
addEpilogue();
}
NeedsEpilogue = false;
for (const auto *Fixup : assembler()->fixups()) {
Fixup->emitOffset(assembler());
}
return AssembledTest(codeBytes(), codeBytesSize(), NumAllocatedDwords);
}
// Allocates a new dword slot in the test's scratchpad area.
uint32_t allocateDword() { return NumAllocatedDwords++; }
// Allocates a new qword slot in the test's scratchpad area.
uint32_t allocateQword() {
uint32_t InitialDword = allocateDword();
allocateDword();
return InitialDword;
}
// Allocates a new dqword slot in the test's scratchpad area.
uint32_t allocateDqword() {
uint32_t InitialDword = allocateQword();
allocateQword();
return InitialDword;
}
Address dwordAddress(uint32_t Dword) {
return Address(Encoded_GPR_r9(), dwordDisp(Dword), nullptr);
}
private:
// e??SlotAddress returns an AssemblerX8664::Traits::Address that can be used
// by the test cases to encode an address operand for accessing the slot for
// the specified register. These are all private for, when jitting the test
// code, tests should not tamper with these values. Besides, during the test
// execution these slots' contents are undefined and should not be accessed.
Address raxSlotAddress() { return dwordAddress(AssembledTest::raxSlot()); }
Address rbxSlotAddress() { return dwordAddress(AssembledTest::rbxSlot()); }
Address rcxSlotAddress() { return dwordAddress(AssembledTest::rcxSlot()); }
Address rdxSlotAddress() { return dwordAddress(AssembledTest::rdxSlot()); }
Address rdiSlotAddress() { return dwordAddress(AssembledTest::rdiSlot()); }
Address rsiSlotAddress() { return dwordAddress(AssembledTest::rsiSlot()); }
Address rbpSlotAddress() { return dwordAddress(AssembledTest::rbpSlot()); }
Address rspSlotAddress() { return dwordAddress(AssembledTest::rspSlot()); }
Address r8SlotAddress() { return dwordAddress(AssembledTest::r8Slot()); }
Address r9SlotAddress() { return dwordAddress(AssembledTest::r9Slot()); }
Address r10SlotAddress() { return dwordAddress(AssembledTest::r10Slot()); }
Address r11SlotAddress() { return dwordAddress(AssembledTest::r11Slot()); }
Address r12SlotAddress() { return dwordAddress(AssembledTest::r12Slot()); }
Address r13SlotAddress() { return dwordAddress(AssembledTest::r13Slot()); }
Address r14SlotAddress() { return dwordAddress(AssembledTest::r14Slot()); }
Address r15SlotAddress() { return dwordAddress(AssembledTest::r15Slot()); }
Address xmm0SlotAddress() { return dwordAddress(AssembledTest::xmm0Slot()); }
Address xmm1SlotAddress() { return dwordAddress(AssembledTest::xmm1Slot()); }
Address xmm2SlotAddress() { return dwordAddress(AssembledTest::xmm2Slot()); }
Address xmm3SlotAddress() { return dwordAddress(AssembledTest::xmm3Slot()); }
Address xmm4SlotAddress() { return dwordAddress(AssembledTest::xmm4Slot()); }
Address xmm5SlotAddress() { return dwordAddress(AssembledTest::xmm5Slot()); }
Address xmm6SlotAddress() { return dwordAddress(AssembledTest::xmm6Slot()); }
Address xmm7SlotAddress() { return dwordAddress(AssembledTest::xmm7Slot()); }
Address xmm8SlotAddress() { return dwordAddress(AssembledTest::xmm8Slot()); }
Address xmm9SlotAddress() { return dwordAddress(AssembledTest::xmm9Slot()); }
Address xmm10SlotAddress() {
return dwordAddress(AssembledTest::xmm10Slot());
}
Address xmm11SlotAddress() {
return dwordAddress(AssembledTest::xmm11Slot());
}
Address xmm12SlotAddress() {
return dwordAddress(AssembledTest::xmm12Slot());
}
Address xmm13SlotAddress() {
return dwordAddress(AssembledTest::xmm13Slot());
}
Address xmm14SlotAddress() {
return dwordAddress(AssembledTest::xmm14Slot());
}
Address xmm15SlotAddress() {
return dwordAddress(AssembledTest::xmm15Slot());
}
// Returns the displacement that should be used when accessing the specified
// Dword in the scratchpad area. It needs to adjust for the initial
// instructions that are emitted before the call that materializes the IP
// register.
uint32_t dwordDisp(uint32_t Dword) const {
EXPECT_LT(Dword, NumAllocatedDwords);
assert(Dword < NumAllocatedDwords);
static constexpr uint8_t PushR9Bytes = 2;
static constexpr uint8_t CallImmBytes = 5;
return AssembledTest::MaximumCodeSize + (Dword * 4) -
(PushR9Bytes + CallImmBytes);
}
void addPrologue() {
__ pushl(Encoded_GPR_r9());
__ call(Immediate(4));
__ popl(Encoded_GPR_r9());
__ pushl(Encoded_GPR_rax());
__ pushl(Encoded_GPR_rbx());
__ pushl(Encoded_GPR_rcx());
__ pushl(Encoded_GPR_rdx());
__ pushl(Encoded_GPR_rbp());
__ pushl(Encoded_GPR_rdi());
__ pushl(Encoded_GPR_rsi());
__ pushl(Encoded_GPR_r8());
__ pushl(Encoded_GPR_r10());
__ pushl(Encoded_GPR_r11());
__ pushl(Encoded_GPR_r12());
__ pushl(Encoded_GPR_r13());
__ pushl(Encoded_GPR_r14());
__ pushl(Encoded_GPR_r15());
__ mov(IceType_i32, Encoded_GPR_rax(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rbx(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rcx(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rdx(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rbp(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rdi(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_rsi(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r8(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r10(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r11(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r12(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r13(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r14(), Immediate(0x00));
__ mov(IceType_i32, Encoded_GPR_r15(), Immediate(0x00));
}
void addEpilogue() {
__ mov(IceType_i64, raxSlotAddress(), Encoded_GPR_rax());
__ mov(IceType_i64, rbxSlotAddress(), Encoded_GPR_rbx());
__ mov(IceType_i64, rcxSlotAddress(), Encoded_GPR_rcx());
__ mov(IceType_i64, rdxSlotAddress(), Encoded_GPR_rdx());
__ mov(IceType_i64, rdiSlotAddress(), Encoded_GPR_rdi());
__ mov(IceType_i64, rsiSlotAddress(), Encoded_GPR_rsi());
__ mov(IceType_i64, rbpSlotAddress(), Encoded_GPR_rbp());
__ mov(IceType_i64, rspSlotAddress(), Encoded_GPR_rsp());
__ mov(IceType_i64, r8SlotAddress(), Encoded_GPR_r8());
__ mov(IceType_i64, r9SlotAddress(), Encoded_GPR_r9());
__ mov(IceType_i64, r10SlotAddress(), Encoded_GPR_r10());
__ mov(IceType_i64, r11SlotAddress(), Encoded_GPR_r11());
__ mov(IceType_i64, r12SlotAddress(), Encoded_GPR_r12());
__ mov(IceType_i64, r13SlotAddress(), Encoded_GPR_r13());
__ mov(IceType_i64, r14SlotAddress(), Encoded_GPR_r14());
__ mov(IceType_i64, r15SlotAddress(), Encoded_GPR_r15());
__ movups(xmm0SlotAddress(), Encoded_Xmm_xmm0());
__ movups(xmm1SlotAddress(), Encoded_Xmm_xmm1());
__ movups(xmm2SlotAddress(), Encoded_Xmm_xmm2());
__ movups(xmm3SlotAddress(), Encoded_Xmm_xmm3());
__ movups(xmm4SlotAddress(), Encoded_Xmm_xmm4());
__ movups(xmm5SlotAddress(), Encoded_Xmm_xmm5());
__ movups(xmm6SlotAddress(), Encoded_Xmm_xmm6());
__ movups(xmm7SlotAddress(), Encoded_Xmm_xmm7());
__ movups(xmm8SlotAddress(), Encoded_Xmm_xmm8());
__ movups(xmm9SlotAddress(), Encoded_Xmm_xmm9());
__ movups(xmm10SlotAddress(), Encoded_Xmm_xmm10());
__ movups(xmm11SlotAddress(), Encoded_Xmm_xmm11());
__ movups(xmm12SlotAddress(), Encoded_Xmm_xmm12());
__ movups(xmm13SlotAddress(), Encoded_Xmm_xmm13());
__ movups(xmm14SlotAddress(), Encoded_Xmm_xmm14());
__ movups(xmm15SlotAddress(), Encoded_Xmm_xmm15());
__ popl(Encoded_GPR_r15());
__ popl(Encoded_GPR_r14());
__ popl(Encoded_GPR_r13());
__ popl(Encoded_GPR_r12());
__ popl(Encoded_GPR_r11());
__ popl(Encoded_GPR_r10());
__ popl(Encoded_GPR_r8());
__ popl(Encoded_GPR_rsi());
__ popl(Encoded_GPR_rdi());
__ popl(Encoded_GPR_rbp());
__ popl(Encoded_GPR_rdx());
__ popl(Encoded_GPR_rcx());
__ popl(Encoded_GPR_rbx());
__ popl(Encoded_GPR_rax());
__ popl(Encoded_GPR_r9());
__ ret();
}
bool NeedsEpilogue;
uint32_t NumAllocatedDwords;
};
} // end of namespace Test
} // end of namespace X8664
} // end of namespace Ice
#endif // ASSEMBLERX8664_TESTUTIL_H_