blob: 27604a88805076c2d80ed7abc182584df72e6811 [file] [log] [blame]
// Copyright 2016 The SwiftShader Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "Reactor.hpp"
#include "Optimizer.hpp"
#include "ExecutableMemory.hpp"
#include "src/IceTypes.h"
#include "src/IceCfg.h"
#include "src/IceELFStreamer.h"
#include "src/IceGlobalContext.h"
#include "src/IceCfgNode.h"
#include "src/IceELFObjectWriter.h"
#include "src/IceGlobalInits.h"
#include "llvm/Support/FileSystem.h"
#include "llvm/Support/raw_os_ostream.h"
#include "llvm/Support/Compiler.h"
#if __has_feature(memory_sanitizer)
#include <sanitizer/msan_interface.h>
#endif
#if defined(_WIN32)
#ifndef WIN32_LEAN_AND_MEAN
#define WIN32_LEAN_AND_MEAN
#endif // !WIN32_LEAN_AND_MEAN
#ifndef NOMINMAX
#define NOMINMAX
#endif // !NOMINMAX
#include <Windows.h>
#else
#include <sys/mman.h>
#if !defined(MAP_ANONYMOUS)
#define MAP_ANONYMOUS MAP_ANON
#endif
#endif
#include <mutex>
#include <limits>
#include <iostream>
#include <cassert>
namespace
{
Ice::GlobalContext *context = nullptr;
Ice::Cfg *function = nullptr;
Ice::CfgNode *basicBlock = nullptr;
Ice::CfgLocalAllocatorScope *allocator = nullptr;
rr::Routine *routine = nullptr;
std::mutex codegenMutex;
Ice::ELFFileStreamer *elfFile = nullptr;
Ice::Fdstream *out = nullptr;
}
namespace
{
#if !defined(__i386__) && defined(_M_IX86)
#define __i386__ 1
#endif
#if !defined(__x86_64__) && (defined(_M_AMD64) || defined (_M_X64))
#define __x86_64__ 1
#endif
class CPUID
{
public:
const static bool ARM;
const static bool SSE4_1;
private:
static void cpuid(int registers[4], int info)
{
#if defined(__i386__) || defined(__x86_64__)
#if defined(_WIN32)
__cpuid(registers, info);
#else
__asm volatile("cpuid": "=a" (registers[0]), "=b" (registers[1]), "=c" (registers[2]), "=d" (registers[3]): "a" (info));
#endif
#else
registers[0] = 0;
registers[1] = 0;
registers[2] = 0;
registers[3] = 0;
#endif
}
static bool detectARM()
{
#if defined(__arm__) || defined(__aarch64__)
return true;
#elif defined(__i386__) || defined(__x86_64__)
return false;
#elif defined(__mips__)
return false;
#else
#error "Unknown architecture"
#endif
}
static bool detectSSE4_1()
{
#if defined(__i386__) || defined(__x86_64__)
int registers[4];
cpuid(registers, 1);
return (registers[2] & 0x00080000) != 0;
#else
return false;
#endif
}
};
const bool CPUID::ARM = CPUID::detectARM();
const bool CPUID::SSE4_1 = CPUID::detectSSE4_1();
const bool emulateIntrinsics = false;
const bool emulateMismatchedBitCast = CPUID::ARM;
}
namespace rr
{
enum EmulatedType
{
EmulatedShift = 16,
EmulatedV2 = 2 << EmulatedShift,
EmulatedV4 = 4 << EmulatedShift,
EmulatedV8 = 8 << EmulatedShift,
EmulatedBits = EmulatedV2 | EmulatedV4 | EmulatedV8,
Type_v2i32 = Ice::IceType_v4i32 | EmulatedV2,
Type_v4i16 = Ice::IceType_v8i16 | EmulatedV4,
Type_v2i16 = Ice::IceType_v8i16 | EmulatedV2,
Type_v8i8 = Ice::IceType_v16i8 | EmulatedV8,
Type_v4i8 = Ice::IceType_v16i8 | EmulatedV4,
Type_v2f32 = Ice::IceType_v4f32 | EmulatedV2,
};
class Value : public Ice::Operand {};
class SwitchCases : public Ice::InstSwitch {};
class BasicBlock : public Ice::CfgNode {};
Ice::Type T(Type *t)
{
static_assert(static_cast<unsigned int>(Ice::IceType_NUM) < static_cast<unsigned int>(EmulatedBits), "Ice::Type overlaps with our emulated types!");
return (Ice::Type)(reinterpret_cast<std::intptr_t>(t) & ~EmulatedBits);
}
Type *T(Ice::Type t)
{
return reinterpret_cast<Type*>(t);
}
Type *T(EmulatedType t)
{
return reinterpret_cast<Type*>(t);
}
Value *V(Ice::Operand *v)
{
return reinterpret_cast<Value*>(v);
}
BasicBlock *B(Ice::CfgNode *b)
{
return reinterpret_cast<BasicBlock*>(b);
}
static size_t typeSize(Type *type)
{
if(reinterpret_cast<std::intptr_t>(type) & EmulatedBits)
{
switch(reinterpret_cast<std::intptr_t>(type))
{
case Type_v2i32: return 8;
case Type_v4i16: return 8;
case Type_v2i16: return 4;
case Type_v8i8: return 8;
case Type_v4i8: return 4;
case Type_v2f32: return 8;
default: assert(false);
}
}
return Ice::typeWidthInBytes(T(type));
}
Optimization optimization[10] = {InstructionCombining, Disabled};
using ElfHeader = std::conditional<sizeof(void*) == 8, Elf64_Ehdr, Elf32_Ehdr>::type;
using SectionHeader = std::conditional<sizeof(void*) == 8, Elf64_Shdr, Elf32_Shdr>::type;
inline const SectionHeader *sectionHeader(const ElfHeader *elfHeader)
{
return reinterpret_cast<const SectionHeader*>((intptr_t)elfHeader + elfHeader->e_shoff);
}
inline const SectionHeader *elfSection(const ElfHeader *elfHeader, int index)
{
return &sectionHeader(elfHeader)[index];
}
static void *relocateSymbol(const ElfHeader *elfHeader, const Elf32_Rel &relocation, const SectionHeader &relocationTable)
{
const SectionHeader *target = elfSection(elfHeader, relocationTable.sh_info);
uint32_t index = relocation.getSymbol();
int table = relocationTable.sh_link;
void *symbolValue = nullptr;
if(index != SHN_UNDEF)
{
if(table == SHN_UNDEF) return nullptr;
const SectionHeader *symbolTable = elfSection(elfHeader, table);
uint32_t symtab_entries = symbolTable->sh_size / symbolTable->sh_entsize;
if(index >= symtab_entries)
{
assert(index < symtab_entries && "Symbol Index out of range");
return nullptr;
}
intptr_t symbolAddress = (intptr_t)elfHeader + symbolTable->sh_offset;
Elf32_Sym &symbol = ((Elf32_Sym*)symbolAddress)[index];
uint16_t section = symbol.st_shndx;
if(section != SHN_UNDEF && section < SHN_LORESERVE)
{
const SectionHeader *target = elfSection(elfHeader, symbol.st_shndx);
symbolValue = reinterpret_cast<void*>((intptr_t)elfHeader + symbol.st_value + target->sh_offset);
}
else
{
return nullptr;
}
}
intptr_t address = (intptr_t)elfHeader + target->sh_offset;
unaligned_ptr<int32_t> patchSite = (int32_t*)(address + relocation.r_offset);
if(CPUID::ARM)
{
switch(relocation.getType())
{
case R_ARM_NONE:
// No relocation
break;
case R_ARM_MOVW_ABS_NC:
{
uint32_t thumb = 0; // Calls to Thumb code not supported.
uint32_t lo = (uint32_t)(intptr_t)symbolValue | thumb;
*patchSite = (*patchSite & 0xFFF0F000) | ((lo & 0xF000) << 4) | (lo & 0x0FFF);
}
break;
case R_ARM_MOVT_ABS:
{
uint32_t hi = (uint32_t)(intptr_t)(symbolValue) >> 16;
*patchSite = (*patchSite & 0xFFF0F000) | ((hi & 0xF000) << 4) | (hi & 0x0FFF);
}
break;
default:
assert(false && "Unsupported relocation type");
return nullptr;
}
}
else
{
switch(relocation.getType())
{
case R_386_NONE:
// No relocation
break;
case R_386_32:
*patchSite = (int32_t)((intptr_t)symbolValue + *patchSite);
break;
// case R_386_PC32:
// *patchSite = (int32_t)((intptr_t)symbolValue + *patchSite - (intptr_t)patchSite);
// break;
default:
assert(false && "Unsupported relocation type");
return nullptr;
}
}
return symbolValue;
}
static void *relocateSymbol(const ElfHeader *elfHeader, const Elf64_Rela &relocation, const SectionHeader &relocationTable)
{
const SectionHeader *target = elfSection(elfHeader, relocationTable.sh_info);
uint32_t index = relocation.getSymbol();
int table = relocationTable.sh_link;
void *symbolValue = nullptr;
if(index != SHN_UNDEF)
{
if(table == SHN_UNDEF) return nullptr;
const SectionHeader *symbolTable = elfSection(elfHeader, table);
uint32_t symtab_entries = symbolTable->sh_size / symbolTable->sh_entsize;
if(index >= symtab_entries)
{
assert(index < symtab_entries && "Symbol Index out of range");
return nullptr;
}
intptr_t symbolAddress = (intptr_t)elfHeader + symbolTable->sh_offset;
Elf64_Sym &symbol = ((Elf64_Sym*)symbolAddress)[index];
uint16_t section = symbol.st_shndx;
if(section != SHN_UNDEF && section < SHN_LORESERVE)
{
const SectionHeader *target = elfSection(elfHeader, symbol.st_shndx);
symbolValue = reinterpret_cast<void*>((intptr_t)elfHeader + symbol.st_value + target->sh_offset);
}
else
{
return nullptr;
}
}
intptr_t address = (intptr_t)elfHeader + target->sh_offset;
unaligned_ptr<int32_t> patchSite32 = (int32_t*)(address + relocation.r_offset);
unaligned_ptr<int64_t> patchSite64 = (int64_t*)(address + relocation.r_offset);
switch(relocation.getType())
{
case R_X86_64_NONE:
// No relocation
break;
case R_X86_64_64:
*patchSite64 = (int64_t)((intptr_t)symbolValue + *patchSite64 + relocation.r_addend);
break;
case R_X86_64_PC32:
*patchSite32 = (int32_t)((intptr_t)symbolValue + *patchSite32 - (intptr_t)patchSite32 + relocation.r_addend);
break;
case R_X86_64_32S:
*patchSite32 = (int32_t)((intptr_t)symbolValue + *patchSite32 + relocation.r_addend);
break;
default:
assert(false && "Unsupported relocation type");
return nullptr;
}
return symbolValue;
}
void *loadImage(uint8_t *const elfImage, size_t &codeSize)
{
ElfHeader *elfHeader = (ElfHeader*)elfImage;
if(!elfHeader->checkMagic())
{
return nullptr;
}
// Expect ELF bitness to match platform
assert(sizeof(void*) == 8 ? elfHeader->getFileClass() == ELFCLASS64 : elfHeader->getFileClass() == ELFCLASS32);
#if defined(__i386__)
assert(sizeof(void*) == 4 && elfHeader->e_machine == EM_386);
#elif defined(__x86_64__)
assert(sizeof(void*) == 8 && elfHeader->e_machine == EM_X86_64);
#elif defined(__arm__)
assert(sizeof(void*) == 4 && elfHeader->e_machine == EM_ARM);
#elif defined(__aarch64__)
assert(sizeof(void*) == 8 && elfHeader->e_machine == EM_AARCH64);
#elif defined(__mips__)
assert(sizeof(void*) == 4 && elfHeader->e_machine == EM_MIPS);
#else
#error "Unsupported platform"
#endif
SectionHeader *sectionHeader = (SectionHeader*)(elfImage + elfHeader->e_shoff);
void *entry = nullptr;
for(int i = 0; i < elfHeader->e_shnum; i++)
{
if(sectionHeader[i].sh_type == SHT_PROGBITS)
{
if(sectionHeader[i].sh_flags & SHF_EXECINSTR)
{
entry = elfImage + sectionHeader[i].sh_offset;
codeSize = sectionHeader[i].sh_size;
}
}
else if(sectionHeader[i].sh_type == SHT_REL)
{
assert(sizeof(void*) == 4 && "UNIMPLEMENTED"); // Only expected/implemented for 32-bit code
for(Elf32_Word index = 0; index < sectionHeader[i].sh_size / sectionHeader[i].sh_entsize; index++)
{
const Elf32_Rel &relocation = ((const Elf32_Rel*)(elfImage + sectionHeader[i].sh_offset))[index];
relocateSymbol(elfHeader, relocation, sectionHeader[i]);
}
}
else if(sectionHeader[i].sh_type == SHT_RELA)
{
assert(sizeof(void*) == 8 && "UNIMPLEMENTED"); // Only expected/implemented for 64-bit code
for(Elf32_Word index = 0; index < sectionHeader[i].sh_size / sectionHeader[i].sh_entsize; index++)
{
const Elf64_Rela &relocation = ((const Elf64_Rela*)(elfImage + sectionHeader[i].sh_offset))[index];
relocateSymbol(elfHeader, relocation, sectionHeader[i]);
}
}
}
return entry;
}
template<typename T>
struct ExecutableAllocator
{
ExecutableAllocator() {};
template<class U> ExecutableAllocator(const ExecutableAllocator<U> &other) {};
using value_type = T;
using size_type = std::size_t;
T *allocate(size_type n)
{
return (T*)allocateExecutable(sizeof(T) * n);
}
void deallocate(T *p, size_type n)
{
deallocateExecutable(p, sizeof(T) * n);
}
};
class ELFMemoryStreamer : public Ice::ELFStreamer, public Routine
{
ELFMemoryStreamer(const ELFMemoryStreamer &) = delete;
ELFMemoryStreamer &operator=(const ELFMemoryStreamer &) = delete;
public:
ELFMemoryStreamer() : Routine(), entry(nullptr)
{
position = 0;
buffer.reserve(0x1000);
}
~ELFMemoryStreamer() override
{
#if defined(_WIN32)
if(buffer.size() != 0)
{
DWORD exeProtection;
VirtualProtect(&buffer[0], buffer.size(), oldProtection, &exeProtection);
}
#endif
}
void write8(uint8_t Value) override
{
if(position == (uint64_t)buffer.size())
{
buffer.push_back(Value);
position++;
}
else if(position < (uint64_t)buffer.size())
{
buffer[position] = Value;
position++;
}
else assert(false && "UNIMPLEMENTED");
}
void writeBytes(llvm::StringRef Bytes) override
{
std::size_t oldSize = buffer.size();
buffer.resize(oldSize + Bytes.size());
memcpy(&buffer[oldSize], Bytes.begin(), Bytes.size());
position += Bytes.size();
}
uint64_t tell() const override { return position; }
void seek(uint64_t Off) override { position = Off; }
const void *getEntry() override
{
if(!entry)
{
position = std::numeric_limits<std::size_t>::max(); // Can't stream more data after this
size_t codeSize = 0;
entry = loadImage(&buffer[0], codeSize);
#if defined(_WIN32)
VirtualProtect(&buffer[0], buffer.size(), PAGE_EXECUTE_READ, &oldProtection);
FlushInstructionCache(GetCurrentProcess(), NULL, 0);
#else
mprotect(&buffer[0], buffer.size(), PROT_READ | PROT_EXEC);
__builtin___clear_cache((char*)entry, (char*)entry + codeSize);
#endif
}
return entry;
}
private:
void *entry;
std::vector<uint8_t, ExecutableAllocator<uint8_t>> buffer;
std::size_t position;
#if defined(_WIN32)
DWORD oldProtection;
#endif
};
Nucleus::Nucleus()
{
::codegenMutex.lock(); // Reactor is currently not thread safe
Ice::ClFlags &Flags = Ice::ClFlags::Flags;
Ice::ClFlags::getParsedClFlags(Flags);
#if defined(__arm__)
Flags.setTargetArch(Ice::Target_ARM32);
Flags.setTargetInstructionSet(Ice::ARM32InstructionSet_HWDivArm);
#elif defined(__mips__)
Flags.setTargetArch(Ice::Target_MIPS32);
Flags.setTargetInstructionSet(Ice::BaseInstructionSet);
#else // x86
Flags.setTargetArch(sizeof(void*) == 8 ? Ice::Target_X8664 : Ice::Target_X8632);
Flags.setTargetInstructionSet(CPUID::SSE4_1 ? Ice::X86InstructionSet_SSE4_1 : Ice::X86InstructionSet_SSE2);
#endif
Flags.setOutFileType(Ice::FT_Elf);
Flags.setOptLevel(Ice::Opt_2);
Flags.setApplicationBinaryInterface(Ice::ABI_Platform);
Flags.setVerbose(false ? Ice::IceV_Most : Ice::IceV_None);
Flags.setDisableHybridAssembly(true);
static llvm::raw_os_ostream cout(std::cout);
static llvm::raw_os_ostream cerr(std::cerr);
if(false) // Write out to a file
{
std::error_code errorCode;
::out = new Ice::Fdstream("out.o", errorCode, llvm::sys::fs::F_None);
::elfFile = new Ice::ELFFileStreamer(*out);
::context = new Ice::GlobalContext(&cout, &cout, &cerr, elfFile);
}
else
{
ELFMemoryStreamer *elfMemory = new ELFMemoryStreamer();
::context = new Ice::GlobalContext(&cout, &cout, &cerr, elfMemory);
::routine = elfMemory;
}
}
Nucleus::~Nucleus()
{
delete ::routine;
delete ::allocator;
delete ::function;
delete ::context;
delete ::elfFile;
delete ::out;
::codegenMutex.unlock();
}
Routine *Nucleus::acquireRoutine(const char *name, bool runOptimizations)
{
if(basicBlock->getInsts().empty() || basicBlock->getInsts().back().getKind() != Ice::Inst::Ret)
{
createRetVoid();
}
::function->setFunctionName(Ice::GlobalString::createWithString(::context, name));
optimize();
::function->translate();
assert(!::function->hasError());
auto globals = ::function->getGlobalInits();
if(globals && !globals->empty())
{
::context->getGlobals()->merge(globals.get());
}
::context->emitFileHeader();
::function->emitIAS();
auto assembler = ::function->releaseAssembler();
auto objectWriter = ::context->getObjectWriter();
assembler->alignFunction();
objectWriter->writeFunctionCode(::function->getFunctionName(), false, assembler.get());
::context->lowerGlobals("last");
::context->lowerConstants();
::context->lowerJumpTables();
objectWriter->setUndefinedSyms(::context->getConstantExternSyms());
objectWriter->writeNonUserSections();
Routine *handoffRoutine = ::routine;
::routine = nullptr;
return handoffRoutine;
}
void Nucleus::optimize()
{
rr::optimize(::function);
}
Value *Nucleus::allocateStackVariable(Type *t, int arraySize)
{
Ice::Type type = T(t);
int typeSize = Ice::typeWidthInBytes(type);
int totalSize = typeSize * (arraySize ? arraySize : 1);
auto bytes = Ice::ConstantInteger32::create(::context, type, totalSize);
auto address = ::function->makeVariable(T(getPointerType(t)));
auto alloca = Ice::InstAlloca::create(::function, address, bytes, typeSize);
::function->getEntryNode()->getInsts().push_front(alloca);
return V(address);
}
BasicBlock *Nucleus::createBasicBlock()
{
return B(::function->makeNode());
}
BasicBlock *Nucleus::getInsertBlock()
{
return B(::basicBlock);
}
void Nucleus::setInsertBlock(BasicBlock *basicBlock)
{
// assert(::basicBlock->getInsts().back().getTerminatorEdges().size() >= 0 && "Previous basic block must have a terminator");
Variable::materializeAll();
::basicBlock = basicBlock;
}
void Nucleus::createFunction(Type *ReturnType, std::vector<Type*> &Params)
{
uint32_t sequenceNumber = 0;
::function = Ice::Cfg::create(::context, sequenceNumber).release();
::allocator = new Ice::CfgLocalAllocatorScope(::function);
for(Type *type : Params)
{
Ice::Variable *arg = ::function->makeVariable(T(type));
::function->addArg(arg);
}
Ice::CfgNode *node = ::function->makeNode();
::function->setEntryNode(node);
::basicBlock = node;
}
Value *Nucleus::getArgument(unsigned int index)
{
return V(::function->getArgs()[index]);
}
void Nucleus::createRetVoid()
{
// Code generated after this point is unreachable, so any variables
// being read can safely return an undefined value. We have to avoid
// materializing variables after the terminator ret instruction.
Variable::killUnmaterialized();
Ice::InstRet *ret = Ice::InstRet::create(::function);
::basicBlock->appendInst(ret);
}
void Nucleus::createRet(Value *v)
{
// Code generated after this point is unreachable, so any variables
// being read can safely return an undefined value. We have to avoid
// materializing variables after the terminator ret instruction.
Variable::killUnmaterialized();
Ice::InstRet *ret = Ice::InstRet::create(::function, v);
::basicBlock->appendInst(ret);
}
void Nucleus::createBr(BasicBlock *dest)
{
Variable::materializeAll();
auto br = Ice::InstBr::create(::function, dest);
::basicBlock->appendInst(br);
}
void Nucleus::createCondBr(Value *cond, BasicBlock *ifTrue, BasicBlock *ifFalse)
{
Variable::materializeAll();
auto br = Ice::InstBr::create(::function, cond, ifTrue, ifFalse);
::basicBlock->appendInst(br);
}
static bool isCommutative(Ice::InstArithmetic::OpKind op)
{
switch(op)
{
case Ice::InstArithmetic::Add:
case Ice::InstArithmetic::Fadd:
case Ice::InstArithmetic::Mul:
case Ice::InstArithmetic::Fmul:
case Ice::InstArithmetic::And:
case Ice::InstArithmetic::Or:
case Ice::InstArithmetic::Xor:
return true;
default:
return false;
}
}
static Value *createArithmetic(Ice::InstArithmetic::OpKind op, Value *lhs, Value *rhs)
{
assert(lhs->getType() == rhs->getType() || llvm::isa<Ice::Constant>(rhs));
bool swapOperands = llvm::isa<Ice::Constant>(lhs) && isCommutative(op);
Ice::Variable *result = ::function->makeVariable(lhs->getType());
Ice::InstArithmetic *arithmetic = Ice::InstArithmetic::create(::function, op, result, swapOperands ? rhs : lhs, swapOperands ? lhs : rhs);
::basicBlock->appendInst(arithmetic);
return V(result);
}
Value *Nucleus::createAdd(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Add, lhs, rhs);
}
Value *Nucleus::createSub(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Sub, lhs, rhs);
}
Value *Nucleus::createMul(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Mul, lhs, rhs);
}
Value *Nucleus::createUDiv(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Udiv, lhs, rhs);
}
Value *Nucleus::createSDiv(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Sdiv, lhs, rhs);
}
Value *Nucleus::createFAdd(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Fadd, lhs, rhs);
}
Value *Nucleus::createFSub(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Fsub, lhs, rhs);
}
Value *Nucleus::createFMul(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Fmul, lhs, rhs);
}
Value *Nucleus::createFDiv(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Fdiv, lhs, rhs);
}
Value *Nucleus::createURem(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Urem, lhs, rhs);
}
Value *Nucleus::createSRem(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Srem, lhs, rhs);
}
Value *Nucleus::createFRem(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Frem, lhs, rhs);
}
Value *Nucleus::createShl(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Shl, lhs, rhs);
}
Value *Nucleus::createLShr(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Lshr, lhs, rhs);
}
Value *Nucleus::createAShr(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Ashr, lhs, rhs);
}
Value *Nucleus::createAnd(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::And, lhs, rhs);
}
Value *Nucleus::createOr(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Or, lhs, rhs);
}
Value *Nucleus::createXor(Value *lhs, Value *rhs)
{
return createArithmetic(Ice::InstArithmetic::Xor, lhs, rhs);
}
Value *Nucleus::createNeg(Value *v)
{
return createSub(createNullValue(T(v->getType())), v);
}
Value *Nucleus::createFNeg(Value *v)
{
double c[4] = {-0.0, -0.0, -0.0, -0.0};
Value *negativeZero = Ice::isVectorType(v->getType()) ?
createConstantVector(c, T(v->getType())) :
V(::context->getConstantFloat(-0.0f));
return createFSub(negativeZero, v);
}
Value *Nucleus::createNot(Value *v)
{
if(Ice::isScalarIntegerType(v->getType()))
{
return createXor(v, V(::context->getConstantInt(v->getType(), -1)));
}
else // Vector
{
int64_t c[16] = {-1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1, -1};
return createXor(v, createConstantVector(c, T(v->getType())));
}
}
Value *Nucleus::createLoad(Value *ptr, Type *type, bool isVolatile, unsigned int align, bool atomic, std::memory_order memoryOrder)
{
assert(!atomic); // Unimplemented
assert(memoryOrder == std::memory_order_relaxed); // Unimplemented
int valueType = (int)reinterpret_cast<intptr_t>(type);
Ice::Variable *result = ::function->makeVariable(T(type));
if((valueType & EmulatedBits) && (align != 0)) // Narrow vector not stored on stack.
{
if(emulateIntrinsics)
{
if(typeSize(type) == 4)
{
auto pointer = RValue<Pointer<Byte>>(ptr);
Int x = *Pointer<Int>(pointer);
Int4 vector;
vector = Insert(vector, x, 0);
auto bitcast = Ice::InstCast::create(::function, Ice::InstCast::Bitcast, result, vector.loadValue());
::basicBlock->appendInst(bitcast);
}
else if(typeSize(type) == 8)
{
auto pointer = RValue<Pointer<Byte>>(ptr);
Int x = *Pointer<Int>(pointer);
Int y = *Pointer<Int>(pointer + 4);
Int4 vector;
vector = Insert(vector, x, 0);
vector = Insert(vector, y, 1);
auto bitcast = Ice::InstCast::create(::function, Ice::InstCast::Bitcast, result, vector.loadValue());
::basicBlock->appendInst(bitcast);
}
else assert(false);
}
else
{
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::LoadSubVector, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto load = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
load->addArg(ptr);
load->addArg(::context->getConstantInt32(typeSize(type)));
::basicBlock->appendInst(load);
}
}
else
{
auto load = Ice::InstLoad::create(::function, result, ptr, align);
::basicBlock->appendInst(load);
}
return V(result);
}
Value *Nucleus::createStore(Value *value, Value *ptr, Type *type, bool isVolatile, unsigned int align, bool atomic, std::memory_order memoryOrder)
{
assert(!atomic); // Unimplemented
assert(memoryOrder == std::memory_order_relaxed); // Unimplemented
#if __has_feature(memory_sanitizer)
// Mark all (non-stack) memory writes as initialized by calling __msan_unpoison
if(align != 0)
{
auto call = Ice::InstCall::create(::function, 2, nullptr, ::context->getConstantInt64(reinterpret_cast<intptr_t>(__msan_unpoison)), false);
call->addArg(ptr);
call->addArg(::context->getConstantInt64(typeSize(type)));
::basicBlock->appendInst(call);
}
#endif
int valueType = (int)reinterpret_cast<intptr_t>(type);
if((valueType & EmulatedBits) && (align != 0)) // Narrow vector not stored on stack.
{
if(emulateIntrinsics)
{
if(typeSize(type) == 4)
{
Ice::Variable *vector = ::function->makeVariable(Ice::IceType_v4i32);
auto bitcast = Ice::InstCast::create(::function, Ice::InstCast::Bitcast, vector, value);
::basicBlock->appendInst(bitcast);
RValue<Int4> v(V(vector));
auto pointer = RValue<Pointer<Byte>>(ptr);
Int x = Extract(v, 0);
*Pointer<Int>(pointer) = x;
}
else if(typeSize(type) == 8)
{
Ice::Variable *vector = ::function->makeVariable(Ice::IceType_v4i32);
auto bitcast = Ice::InstCast::create(::function, Ice::InstCast::Bitcast, vector, value);
::basicBlock->appendInst(bitcast);
RValue<Int4> v(V(vector));
auto pointer = RValue<Pointer<Byte>>(ptr);
Int x = Extract(v, 0);
*Pointer<Int>(pointer) = x;
Int y = Extract(v, 1);
*Pointer<Int>(pointer + 4) = y;
}
else assert(false);
}
else
{
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::StoreSubVector, Ice::Intrinsics::SideEffects_T, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_T};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto store = Ice::InstIntrinsicCall::create(::function, 3, nullptr, target, intrinsic);
store->addArg(value);
store->addArg(ptr);
store->addArg(::context->getConstantInt32(typeSize(type)));
::basicBlock->appendInst(store);
}
}
else
{
assert(value->getType() == T(type));
auto store = Ice::InstStore::create(::function, value, ptr, align);
::basicBlock->appendInst(store);
}
return value;
}
Value *Nucleus::createGEP(Value *ptr, Type *type, Value *index, bool unsignedIndex)
{
assert(index->getType() == Ice::IceType_i32);
if(auto *constant = llvm::dyn_cast<Ice::ConstantInteger32>(index))
{
int32_t offset = constant->getValue() * (int)typeSize(type);
if(offset == 0)
{
return ptr;
}
return createAdd(ptr, createConstantInt(offset));
}
if(!Ice::isByteSizedType(T(type)))
{
index = createMul(index, createConstantInt((int)typeSize(type)));
}
if(sizeof(void*) == 8)
{
if(unsignedIndex)
{
index = createZExt(index, T(Ice::IceType_i64));
}
else
{
index = createSExt(index, T(Ice::IceType_i64));
}
}
return createAdd(ptr, index);
}
Value *Nucleus::createAtomicAdd(Value *ptr, Value *value)
{
assert(false && "UNIMPLEMENTED"); return nullptr;
}
static Value *createCast(Ice::InstCast::OpKind op, Value *v, Type *destType)
{
if(v->getType() == T(destType))
{
return v;
}
Ice::Variable *result = ::function->makeVariable(T(destType));
Ice::InstCast *cast = Ice::InstCast::create(::function, op, result, v);
::basicBlock->appendInst(cast);
return V(result);
}
Value *Nucleus::createTrunc(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Trunc, v, destType);
}
Value *Nucleus::createZExt(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Zext, v, destType);
}
Value *Nucleus::createSExt(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Sext, v, destType);
}
Value *Nucleus::createFPToSI(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Fptosi, v, destType);
}
Value *Nucleus::createSIToFP(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Sitofp, v, destType);
}
Value *Nucleus::createFPTrunc(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Fptrunc, v, destType);
}
Value *Nucleus::createFPExt(Value *v, Type *destType)
{
return createCast(Ice::InstCast::Fpext, v, destType);
}
Value *Nucleus::createBitCast(Value *v, Type *destType)
{
// Bitcasts must be between types of the same logical size. But with emulated narrow vectors we need
// support for casting between scalars and wide vectors. For platforms where this is not supported,
// emulate them by writing to the stack and reading back as the destination type.
if(emulateMismatchedBitCast)
{
if(!Ice::isVectorType(v->getType()) && Ice::isVectorType(T(destType)))
{
Value *address = allocateStackVariable(destType);
createStore(v, address, T(v->getType()));
return createLoad(address, destType);
}
else if(Ice::isVectorType(v->getType()) && !Ice::isVectorType(T(destType)))
{
Value *address = allocateStackVariable(T(v->getType()));
createStore(v, address, T(v->getType()));
return createLoad(address, destType);
}
}
return createCast(Ice::InstCast::Bitcast, v, destType);
}
static Value *createIntCompare(Ice::InstIcmp::ICond condition, Value *lhs, Value *rhs)
{
assert(lhs->getType() == rhs->getType());
auto result = ::function->makeVariable(Ice::isScalarIntegerType(lhs->getType()) ? Ice::IceType_i1 : lhs->getType());
auto cmp = Ice::InstIcmp::create(::function, condition, result, lhs, rhs);
::basicBlock->appendInst(cmp);
return V(result);
}
Value *Nucleus::createICmpEQ(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Eq, lhs, rhs);
}
Value *Nucleus::createICmpNE(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Ne, lhs, rhs);
}
Value *Nucleus::createICmpUGT(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Ugt, lhs, rhs);
}
Value *Nucleus::createICmpUGE(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Uge, lhs, rhs);
}
Value *Nucleus::createICmpULT(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Ult, lhs, rhs);
}
Value *Nucleus::createICmpULE(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Ule, lhs, rhs);
}
Value *Nucleus::createICmpSGT(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Sgt, lhs, rhs);
}
Value *Nucleus::createICmpSGE(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Sge, lhs, rhs);
}
Value *Nucleus::createICmpSLT(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Slt, lhs, rhs);
}
Value *Nucleus::createICmpSLE(Value *lhs, Value *rhs)
{
return createIntCompare(Ice::InstIcmp::Sle, lhs, rhs);
}
static Value *createFloatCompare(Ice::InstFcmp::FCond condition, Value *lhs, Value *rhs)
{
assert(lhs->getType() == rhs->getType());
assert(Ice::isScalarFloatingType(lhs->getType()) || lhs->getType() == Ice::IceType_v4f32);
auto result = ::function->makeVariable(Ice::isScalarFloatingType(lhs->getType()) ? Ice::IceType_i1 : Ice::IceType_v4i32);
auto cmp = Ice::InstFcmp::create(::function, condition, result, lhs, rhs);
::basicBlock->appendInst(cmp);
return V(result);
}
Value *Nucleus::createFCmpOEQ(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Oeq, lhs, rhs);
}
Value *Nucleus::createFCmpOGT(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ogt, lhs, rhs);
}
Value *Nucleus::createFCmpOGE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Oge, lhs, rhs);
}
Value *Nucleus::createFCmpOLT(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Olt, lhs, rhs);
}
Value *Nucleus::createFCmpOLE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ole, lhs, rhs);
}
Value *Nucleus::createFCmpONE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::One, lhs, rhs);
}
Value *Nucleus::createFCmpORD(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ord, lhs, rhs);
}
Value *Nucleus::createFCmpUNO(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Uno, lhs, rhs);
}
Value *Nucleus::createFCmpUEQ(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ueq, lhs, rhs);
}
Value *Nucleus::createFCmpUGT(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ugt, lhs, rhs);
}
Value *Nucleus::createFCmpUGE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Uge, lhs, rhs);
}
Value *Nucleus::createFCmpULT(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ult, lhs, rhs);
}
Value *Nucleus::createFCmpULE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Ule, lhs, rhs);
}
Value *Nucleus::createFCmpUNE(Value *lhs, Value *rhs)
{
return createFloatCompare(Ice::InstFcmp::Une, lhs, rhs);
}
Value *Nucleus::createExtractElement(Value *vector, Type *type, int index)
{
auto result = ::function->makeVariable(T(type));
auto extract = Ice::InstExtractElement::create(::function, result, vector, ::context->getConstantInt32(index));
::basicBlock->appendInst(extract);
return V(result);
}
Value *Nucleus::createInsertElement(Value *vector, Value *element, int index)
{
auto result = ::function->makeVariable(vector->getType());
auto insert = Ice::InstInsertElement::create(::function, result, vector, element, ::context->getConstantInt32(index));
::basicBlock->appendInst(insert);
return V(result);
}
Value *Nucleus::createShuffleVector(Value *V1, Value *V2, const int *select)
{
assert(V1->getType() == V2->getType());
int size = Ice::typeNumElements(V1->getType());
auto result = ::function->makeVariable(V1->getType());
auto shuffle = Ice::InstShuffleVector::create(::function, result, V1, V2);
for(int i = 0; i < size; i++)
{
shuffle->addIndex(llvm::cast<Ice::ConstantInteger32>(::context->getConstantInt32(select[i])));
}
::basicBlock->appendInst(shuffle);
return V(result);
}
Value *Nucleus::createSelect(Value *C, Value *ifTrue, Value *ifFalse)
{
assert(ifTrue->getType() == ifFalse->getType());
auto result = ::function->makeVariable(ifTrue->getType());
auto *select = Ice::InstSelect::create(::function, result, C, ifTrue, ifFalse);
::basicBlock->appendInst(select);
return V(result);
}
SwitchCases *Nucleus::createSwitch(Value *control, BasicBlock *defaultBranch, unsigned numCases)
{
auto switchInst = Ice::InstSwitch::create(::function, numCases, control, defaultBranch);
::basicBlock->appendInst(switchInst);
return reinterpret_cast<SwitchCases*>(switchInst);
}
void Nucleus::addSwitchCase(SwitchCases *switchCases, int label, BasicBlock *branch)
{
switchCases->addBranch(label, label, branch);
}
void Nucleus::createUnreachable()
{
Ice::InstUnreachable *unreachable = Ice::InstUnreachable::create(::function);
::basicBlock->appendInst(unreachable);
}
Type *Nucleus::getPointerType(Type *ElementType)
{
if(sizeof(void*) == 8)
{
return T(Ice::IceType_i64);
}
else
{
return T(Ice::IceType_i32);
}
}
Value *Nucleus::createNullValue(Type *Ty)
{
if(Ice::isVectorType(T(Ty)))
{
assert(Ice::typeNumElements(T(Ty)) <= 16);
int64_t c[16] = {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0};
return createConstantVector(c, Ty);
}
else
{
return V(::context->getConstantZero(T(Ty)));
}
}
Value *Nucleus::createConstantLong(int64_t i)
{
return V(::context->getConstantInt64(i));
}
Value *Nucleus::createConstantInt(int i)
{
return V(::context->getConstantInt32(i));
}
Value *Nucleus::createConstantInt(unsigned int i)
{
return V(::context->getConstantInt32(i));
}
Value *Nucleus::createConstantBool(bool b)
{
return V(::context->getConstantInt1(b));
}
Value *Nucleus::createConstantByte(signed char i)
{
return V(::context->getConstantInt8(i));
}
Value *Nucleus::createConstantByte(unsigned char i)
{
return V(::context->getConstantInt8(i));
}
Value *Nucleus::createConstantShort(short i)
{
return V(::context->getConstantInt16(i));
}
Value *Nucleus::createConstantShort(unsigned short i)
{
return V(::context->getConstantInt16(i));
}
Value *Nucleus::createConstantFloat(float x)
{
return V(::context->getConstantFloat(x));
}
Value *Nucleus::createNullPointer(Type *Ty)
{
return createNullValue(T(sizeof(void*) == 8 ? Ice::IceType_i64 : Ice::IceType_i32));
}
Value *Nucleus::createConstantVector(const int64_t *constants, Type *type)
{
const int vectorSize = 16;
assert(Ice::typeWidthInBytes(T(type)) == vectorSize);
const int alignment = vectorSize;
auto globalPool = ::function->getGlobalPool();
const int64_t *i = constants;
const double *f = reinterpret_cast<const double*>(constants);
Ice::VariableDeclaration::DataInitializer *dataInitializer = nullptr;
switch((int)reinterpret_cast<intptr_t>(type))
{
case Ice::IceType_v4i32:
case Ice::IceType_v4i1:
{
const int initializer[4] = {(int)i[0], (int)i[1], (int)i[2], (int)i[3]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Ice::IceType_v4f32:
{
const float initializer[4] = {(float)f[0], (float)f[1], (float)f[2], (float)f[3]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Ice::IceType_v8i16:
case Ice::IceType_v8i1:
{
const short initializer[8] = {(short)i[0], (short)i[1], (short)i[2], (short)i[3], (short)i[4], (short)i[5], (short)i[6], (short)i[7]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Ice::IceType_v16i8:
case Ice::IceType_v16i1:
{
const char initializer[16] = {(char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[4], (char)i[5], (char)i[6], (char)i[7], (char)i[8], (char)i[9], (char)i[10], (char)i[11], (char)i[12], (char)i[13], (char)i[14], (char)i[15]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Type_v2i32:
{
const int initializer[4] = {(int)i[0], (int)i[1], (int)i[0], (int)i[1]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Type_v2f32:
{
const float initializer[4] = {(float)f[0], (float)f[1], (float)f[0], (float)f[1]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Type_v4i16:
{
const short initializer[8] = {(short)i[0], (short)i[1], (short)i[2], (short)i[3], (short)i[0], (short)i[1], (short)i[2], (short)i[3]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Type_v8i8:
{
const char initializer[16] = {(char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[4], (char)i[5], (char)i[6], (char)i[7], (char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[4], (char)i[5], (char)i[6], (char)i[7]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
case Type_v4i8:
{
const char initializer[16] = {(char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[0], (char)i[1], (char)i[2], (char)i[3], (char)i[0], (char)i[1], (char)i[2], (char)i[3]};
static_assert(sizeof(initializer) == vectorSize, "!");
dataInitializer = Ice::VariableDeclaration::DataInitializer::create(globalPool, (const char*)initializer, vectorSize);
}
break;
default:
assert(false && "Unknown constant vector type" && type);
}
auto name = Ice::GlobalString::createWithoutString(::context);
auto *variableDeclaration = Ice::VariableDeclaration::create(globalPool);
variableDeclaration->setName(name);
variableDeclaration->setAlignment(alignment);
variableDeclaration->setIsConstant(true);
variableDeclaration->addInitializer(dataInitializer);
::function->addGlobal(variableDeclaration);
constexpr int32_t offset = 0;
Ice::Operand *ptr = ::context->getConstantSym(offset, name);
Ice::Variable *result = ::function->makeVariable(T(type));
auto load = Ice::InstLoad::create(::function, result, ptr, alignment);
::basicBlock->appendInst(load);
return V(result);
}
Value *Nucleus::createConstantVector(const double *constants, Type *type)
{
return createConstantVector((const int64_t*)constants, type);
}
Type *Void::getType()
{
return T(Ice::IceType_void);
}
Type *Bool::getType()
{
return T(Ice::IceType_i1);
}
Type *Byte::getType()
{
return T(Ice::IceType_i8);
}
Type *SByte::getType()
{
return T(Ice::IceType_i8);
}
Type *Short::getType()
{
return T(Ice::IceType_i16);
}
Type *UShort::getType()
{
return T(Ice::IceType_i16);
}
Type *Byte4::getType()
{
return T(Type_v4i8);
}
Type *SByte4::getType()
{
return T(Type_v4i8);
}
namespace
{
RValue<Byte> SaturateUnsigned(RValue<Short> x)
{
return Byte(IfThenElse(Int(x) > 0xFF, Int(0xFF), IfThenElse(Int(x) < 0, Int(0), Int(x))));
}
RValue<Byte> Extract(RValue<Byte8> val, int i)
{
return RValue<Byte>(Nucleus::createExtractElement(val.value, Byte::getType(), i));
}
RValue<Byte8> Insert(RValue<Byte8> val, RValue<Byte> element, int i)
{
return RValue<Byte8>(Nucleus::createInsertElement(val.value, element.value, i));
}
}
RValue<Byte8> AddSat(RValue<Byte8> x, RValue<Byte8> y)
{
if(emulateIntrinsics)
{
Byte8 result;
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 0)) + Int(Extract(y, 0)))), 0);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 1)) + Int(Extract(y, 1)))), 1);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 2)) + Int(Extract(y, 2)))), 2);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 3)) + Int(Extract(y, 3)))), 3);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 4)) + Int(Extract(y, 4)))), 4);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 5)) + Int(Extract(y, 5)))), 5);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 6)) + Int(Extract(y, 6)))), 6);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 7)) + Int(Extract(y, 7)))), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::AddSaturateUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto paddusb = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
paddusb->addArg(x.value);
paddusb->addArg(y.value);
::basicBlock->appendInst(paddusb);
return RValue<Byte8>(V(result));
}
}
RValue<Byte8> SubSat(RValue<Byte8> x, RValue<Byte8> y)
{
if(emulateIntrinsics)
{
Byte8 result;
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 0)) - Int(Extract(y, 0)))), 0);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 1)) - Int(Extract(y, 1)))), 1);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 2)) - Int(Extract(y, 2)))), 2);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 3)) - Int(Extract(y, 3)))), 3);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 4)) - Int(Extract(y, 4)))), 4);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 5)) - Int(Extract(y, 5)))), 5);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 6)) - Int(Extract(y, 6)))), 6);
result = Insert(result, SaturateUnsigned(Short(Int(Extract(x, 7)) - Int(Extract(y, 7)))), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SubtractSaturateUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto psubusw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
psubusw->addArg(x.value);
psubusw->addArg(y.value);
::basicBlock->appendInst(psubusw);
return RValue<Byte8>(V(result));
}
}
RValue<SByte> Extract(RValue<SByte8> val, int i)
{
return RValue<SByte>(Nucleus::createExtractElement(val.value, SByte::getType(), i));
}
RValue<SByte8> Insert(RValue<SByte8> val, RValue<SByte> element, int i)
{
return RValue<SByte8>(Nucleus::createInsertElement(val.value, element.value, i));
}
RValue<SByte8> operator>>(RValue<SByte8> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
SByte8 result;
result = Insert(result, Extract(lhs, 0) >> SByte(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> SByte(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> SByte(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> SByte(rhs), 3);
result = Insert(result, Extract(lhs, 4) >> SByte(rhs), 4);
result = Insert(result, Extract(lhs, 5) >> SByte(rhs), 5);
result = Insert(result, Extract(lhs, 6) >> SByte(rhs), 6);
result = Insert(result, Extract(lhs, 7) >> SByte(rhs), 7);
return result;
}
else
{
#if defined(__i386__) || defined(__x86_64__)
// SSE2 doesn't support byte vector shifts, so shift as shorts and recombine.
RValue<Short4> hi = (As<Short4>(lhs) >> rhs) & Short4(0xFF00u);
RValue<Short4> lo = As<Short4>(As<UShort4>((As<Short4>(lhs) << 8) >> rhs) >> 8);
return As<SByte8>(hi | lo);
#else
return RValue<SByte8>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
#endif
}
}
RValue<Int> SignMask(RValue<Byte8> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
Byte8 xx = As<Byte8>(As<SByte8>(x) >> 7) & Byte8(0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80);
return Int(Extract(xx, 0)) | Int(Extract(xx, 1)) | Int(Extract(xx, 2)) | Int(Extract(xx, 3)) | Int(Extract(xx, 4)) | Int(Extract(xx, 5)) | Int(Extract(xx, 6)) | Int(Extract(xx, 7));
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SignMask, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto movmsk = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
movmsk->addArg(x.value);
::basicBlock->appendInst(movmsk);
return RValue<Int>(V(result)) & 0xFF;
}
}
// RValue<Byte8> CmpGT(RValue<Byte8> x, RValue<Byte8> y)
// {
// return RValue<Byte8>(createIntCompare(Ice::InstIcmp::Ugt, x.value, y.value));
// }
RValue<Byte8> CmpEQ(RValue<Byte8> x, RValue<Byte8> y)
{
return RValue<Byte8>(Nucleus::createICmpEQ(x.value, y.value));
}
Type *Byte8::getType()
{
return T(Type_v8i8);
}
// RValue<SByte8> operator<<(RValue<SByte8> lhs, unsigned char rhs)
// {
// return RValue<SByte8>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
// }
// RValue<SByte8> operator>>(RValue<SByte8> lhs, unsigned char rhs)
// {
// return RValue<SByte8>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
// }
RValue<SByte> SaturateSigned(RValue<Short> x)
{
return SByte(IfThenElse(Int(x) > 0x7F, Int(0x7F), IfThenElse(Int(x) < -0x80, Int(0x80), Int(x))));
}
RValue<SByte8> AddSat(RValue<SByte8> x, RValue<SByte8> y)
{
if(emulateIntrinsics)
{
SByte8 result;
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 0)) + Int(Extract(y, 0)))), 0);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 1)) + Int(Extract(y, 1)))), 1);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 2)) + Int(Extract(y, 2)))), 2);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 3)) + Int(Extract(y, 3)))), 3);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 4)) + Int(Extract(y, 4)))), 4);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 5)) + Int(Extract(y, 5)))), 5);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 6)) + Int(Extract(y, 6)))), 6);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 7)) + Int(Extract(y, 7)))), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::AddSaturateSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto paddsb = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
paddsb->addArg(x.value);
paddsb->addArg(y.value);
::basicBlock->appendInst(paddsb);
return RValue<SByte8>(V(result));
}
}
RValue<SByte8> SubSat(RValue<SByte8> x, RValue<SByte8> y)
{
if(emulateIntrinsics)
{
SByte8 result;
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 0)) - Int(Extract(y, 0)))), 0);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 1)) - Int(Extract(y, 1)))), 1);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 2)) - Int(Extract(y, 2)))), 2);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 3)) - Int(Extract(y, 3)))), 3);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 4)) - Int(Extract(y, 4)))), 4);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 5)) - Int(Extract(y, 5)))), 5);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 6)) - Int(Extract(y, 6)))), 6);
result = Insert(result, SaturateSigned(Short(Int(Extract(x, 7)) - Int(Extract(y, 7)))), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SubtractSaturateSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto psubsb = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
psubsb->addArg(x.value);
psubsb->addArg(y.value);
::basicBlock->appendInst(psubsb);
return RValue<SByte8>(V(result));
}
}
RValue<Int> SignMask(RValue<SByte8> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
SByte8 xx = (x >> 7) & SByte8(0x01, 0x02, 0x04, 0x08, 0x10, 0x20, 0x40, 0x80);
return Int(Extract(xx, 0)) | Int(Extract(xx, 1)) | Int(Extract(xx, 2)) | Int(Extract(xx, 3)) | Int(Extract(xx, 4)) | Int(Extract(xx, 5)) | Int(Extract(xx, 6)) | Int(Extract(xx, 7));
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SignMask, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto movmsk = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
movmsk->addArg(x.value);
::basicBlock->appendInst(movmsk);
return RValue<Int>(V(result)) & 0xFF;
}
}
RValue<Byte8> CmpGT(RValue<SByte8> x, RValue<SByte8> y)
{
return RValue<Byte8>(createIntCompare(Ice::InstIcmp::Sgt, x.value, y.value));
}
RValue<Byte8> CmpEQ(RValue<SByte8> x, RValue<SByte8> y)
{
return RValue<Byte8>(Nucleus::createICmpEQ(x.value, y.value));
}
Type *SByte8::getType()
{
return T(Type_v8i8);
}
Type *Byte16::getType()
{
return T(Ice::IceType_v16i8);
}
Type *SByte16::getType()
{
return T(Ice::IceType_v16i8);
}
Type *Short2::getType()
{
return T(Type_v2i16);
}
Type *UShort2::getType()
{
return T(Type_v2i16);
}
Short4::Short4(RValue<Int4> cast)
{
int select[8] = {0, 2, 4, 6, 0, 2, 4, 6};
Value *short8 = Nucleus::createBitCast(cast.value, Short8::getType());
Value *packed = Nucleus::createShuffleVector(short8, short8, select);
Value *int2 = RValue<Int2>(Int2(As<Int4>(packed))).value;
Value *short4 = Nucleus::createBitCast(int2, Short4::getType());
storeValue(short4);
}
// Short4::Short4(RValue<Float> cast)
// {
// }
Short4::Short4(RValue<Float4> cast)
{
assert(false && "UNIMPLEMENTED");
}
RValue<Short4> operator<<(RValue<Short4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Short4 result;
result = Insert(result, Extract(lhs, 0) << Short(rhs), 0);
result = Insert(result, Extract(lhs, 1) << Short(rhs), 1);
result = Insert(result, Extract(lhs, 2) << Short(rhs), 2);
result = Insert(result, Extract(lhs, 3) << Short(rhs), 3);
return result;
}
else
{
return RValue<Short4>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Short4> operator>>(RValue<Short4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Short4 result;
result = Insert(result, Extract(lhs, 0) >> Short(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> Short(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> Short(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> Short(rhs), 3);
return result;
}
else
{
return RValue<Short4>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Short4> Max(RValue<Short4> x, RValue<Short4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v8i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Sle, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<Short4>(V(result));
}
RValue<Short4> Min(RValue<Short4> x, RValue<Short4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v8i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Sgt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<Short4>(V(result));
}
RValue<Short> SaturateSigned(RValue<Int> x)
{
return Short(IfThenElse(x > 0x7FFF, Int(0x7FFF), IfThenElse(x < -0x8000, Int(0x8000), x)));
}
RValue<Short4> AddSat(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
Short4 result;
result = Insert(result, SaturateSigned(Int(Extract(x, 0)) + Int(Extract(y, 0))), 0);
result = Insert(result, SaturateSigned(Int(Extract(x, 1)) + Int(Extract(y, 1))), 1);
result = Insert(result, SaturateSigned(Int(Extract(x, 2)) + Int(Extract(y, 2))), 2);
result = Insert(result, SaturateSigned(Int(Extract(x, 3)) + Int(Extract(y, 3))), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::AddSaturateSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto paddsw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
paddsw->addArg(x.value);
paddsw->addArg(y.value);
::basicBlock->appendInst(paddsw);
return RValue<Short4>(V(result));
}
}
RValue<Short4> SubSat(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
Short4 result;
result = Insert(result, SaturateSigned(Int(Extract(x, 0)) - Int(Extract(y, 0))), 0);
result = Insert(result, SaturateSigned(Int(Extract(x, 1)) - Int(Extract(y, 1))), 1);
result = Insert(result, SaturateSigned(Int(Extract(x, 2)) - Int(Extract(y, 2))), 2);
result = Insert(result, SaturateSigned(Int(Extract(x, 3)) - Int(Extract(y, 3))), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SubtractSaturateSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto psubsw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
psubsw->addArg(x.value);
psubsw->addArg(y.value);
::basicBlock->appendInst(psubsw);
return RValue<Short4>(V(result));
}
}
RValue<Short4> MulHigh(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
Short4 result;
result = Insert(result, Short((Int(Extract(x, 0)) * Int(Extract(y, 0))) >> 16), 0);
result = Insert(result, Short((Int(Extract(x, 1)) * Int(Extract(y, 1))) >> 16), 1);
result = Insert(result, Short((Int(Extract(x, 2)) * Int(Extract(y, 2))) >> 16), 2);
result = Insert(result, Short((Int(Extract(x, 3)) * Int(Extract(y, 3))) >> 16), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::MultiplyHighSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pmulhw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pmulhw->addArg(x.value);
pmulhw->addArg(y.value);
::basicBlock->appendInst(pmulhw);
return RValue<Short4>(V(result));
}
}
RValue<Int2> MulAdd(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
Int2 result;
result = Insert(result, Int(Extract(x, 0)) * Int(Extract(y, 0)) + Int(Extract(x, 1)) * Int(Extract(y, 1)), 0);
result = Insert(result, Int(Extract(x, 2)) * Int(Extract(y, 2)) + Int(Extract(x, 3)) * Int(Extract(y, 3)), 1);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::MultiplyAddPairs, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pmaddwd = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pmaddwd->addArg(x.value);
pmaddwd->addArg(y.value);
::basicBlock->appendInst(pmaddwd);
return As<Int2>(V(result));
}
}
RValue<SByte8> PackSigned(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
SByte8 result;
result = Insert(result, SaturateSigned(Extract(x, 0)), 0);
result = Insert(result, SaturateSigned(Extract(x, 1)), 1);
result = Insert(result, SaturateSigned(Extract(x, 2)), 2);
result = Insert(result, SaturateSigned(Extract(x, 3)), 3);
result = Insert(result, SaturateSigned(Extract(y, 0)), 4);
result = Insert(result, SaturateSigned(Extract(y, 1)), 5);
result = Insert(result, SaturateSigned(Extract(y, 2)), 6);
result = Insert(result, SaturateSigned(Extract(y, 3)), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::VectorPackSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pack = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pack->addArg(x.value);
pack->addArg(y.value);
::basicBlock->appendInst(pack);
return As<SByte8>(Swizzle(As<Int4>(V(result)), 0x88));
}
}
RValue<Byte8> PackUnsigned(RValue<Short4> x, RValue<Short4> y)
{
if(emulateIntrinsics)
{
Byte8 result;
result = Insert(result, SaturateUnsigned(Extract(x, 0)), 0);
result = Insert(result, SaturateUnsigned(Extract(x, 1)), 1);
result = Insert(result, SaturateUnsigned(Extract(x, 2)), 2);
result = Insert(result, SaturateUnsigned(Extract(x, 3)), 3);
result = Insert(result, SaturateUnsigned(Extract(y, 0)), 4);
result = Insert(result, SaturateUnsigned(Extract(y, 1)), 5);
result = Insert(result, SaturateUnsigned(Extract(y, 2)), 6);
result = Insert(result, SaturateUnsigned(Extract(y, 3)), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v16i8);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::VectorPackUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pack = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pack->addArg(x.value);
pack->addArg(y.value);
::basicBlock->appendInst(pack);
return As<Byte8>(Swizzle(As<Int4>(V(result)), 0x88));
}
}
RValue<Short4> CmpGT(RValue<Short4> x, RValue<Short4> y)
{
return RValue<Short4>(createIntCompare(Ice::InstIcmp::Sgt, x.value, y.value));
}
RValue<Short4> CmpEQ(RValue<Short4> x, RValue<Short4> y)
{
return RValue<Short4>(Nucleus::createICmpEQ(x.value, y.value));
}
Type *Short4::getType()
{
return T(Type_v4i16);
}
UShort4::UShort4(RValue<Float4> cast, bool saturate)
{
if(saturate)
{
if(CPUID::SSE4_1)
{
// x86 produces 0x80000000 on 32-bit integer overflow/underflow.
// PackUnsigned takes care of 0x0000 saturation.
Int4 int4(Min(cast, Float4(0xFFFF)));
*this = As<UShort4>(PackUnsigned(int4, int4));
}
else if(CPUID::ARM)
{
// ARM saturates the 32-bit integer result on overflow/undeflow.
Int4 int4(cast);
*this = As<UShort4>(PackUnsigned(int4, int4));
}
else
{
*this = Short4(Int4(Max(Min(cast, Float4(0xFFFF)), Float4(0x0000))));
}
}
else
{
*this = Short4(Int4(cast));
}
}
RValue<UShort> Extract(RValue<UShort4> val, int i)
{
return RValue<UShort>(Nucleus::createExtractElement(val.value, UShort::getType(), i));
}
RValue<UShort4> Insert(RValue<UShort4> val, RValue<UShort> element, int i)
{
return RValue<UShort4>(Nucleus::createInsertElement(val.value, element.value, i));
}
RValue<UShort4> operator<<(RValue<UShort4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UShort4 result;
result = Insert(result, Extract(lhs, 0) << UShort(rhs), 0);
result = Insert(result, Extract(lhs, 1) << UShort(rhs), 1);
result = Insert(result, Extract(lhs, 2) << UShort(rhs), 2);
result = Insert(result, Extract(lhs, 3) << UShort(rhs), 3);
return result;
}
else
{
return RValue<UShort4>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UShort4> operator>>(RValue<UShort4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UShort4 result;
result = Insert(result, Extract(lhs, 0) >> UShort(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> UShort(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> UShort(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> UShort(rhs), 3);
return result;
}
else
{
return RValue<UShort4>(Nucleus::createLShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UShort4> Max(RValue<UShort4> x, RValue<UShort4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v8i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Ule, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<UShort4>(V(result));
}
RValue<UShort4> Min(RValue<UShort4> x, RValue<UShort4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v8i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Ugt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<UShort4>(V(result));
}
RValue<UShort> SaturateUnsigned(RValue<Int> x)
{
return UShort(IfThenElse(x > 0xFFFF, Int(0xFFFF), IfThenElse(x < 0, Int(0), x)));
}
RValue<UShort4> AddSat(RValue<UShort4> x, RValue<UShort4> y)
{
if(emulateIntrinsics)
{
UShort4 result;
result = Insert(result, SaturateUnsigned(Int(Extract(x, 0)) + Int(Extract(y, 0))), 0);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 1)) + Int(Extract(y, 1))), 1);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 2)) + Int(Extract(y, 2))), 2);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 3)) + Int(Extract(y, 3))), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::AddSaturateUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto paddusw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
paddusw->addArg(x.value);
paddusw->addArg(y.value);
::basicBlock->appendInst(paddusw);
return RValue<UShort4>(V(result));
}
}
RValue<UShort4> SubSat(RValue<UShort4> x, RValue<UShort4> y)
{
if(emulateIntrinsics)
{
UShort4 result;
result = Insert(result, SaturateUnsigned(Int(Extract(x, 0)) - Int(Extract(y, 0))), 0);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 1)) - Int(Extract(y, 1))), 1);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 2)) - Int(Extract(y, 2))), 2);
result = Insert(result, SaturateUnsigned(Int(Extract(x, 3)) - Int(Extract(y, 3))), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SubtractSaturateUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto psubusw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
psubusw->addArg(x.value);
psubusw->addArg(y.value);
::basicBlock->appendInst(psubusw);
return RValue<UShort4>(V(result));
}
}
RValue<UShort4> MulHigh(RValue<UShort4> x, RValue<UShort4> y)
{
if(emulateIntrinsics)
{
UShort4 result;
result = Insert(result, UShort((UInt(Extract(x, 0)) * UInt(Extract(y, 0))) >> 16), 0);
result = Insert(result, UShort((UInt(Extract(x, 1)) * UInt(Extract(y, 1))) >> 16), 1);
result = Insert(result, UShort((UInt(Extract(x, 2)) * UInt(Extract(y, 2))) >> 16), 2);
result = Insert(result, UShort((UInt(Extract(x, 3)) * UInt(Extract(y, 3))) >> 16), 3);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::MultiplyHighUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pmulhuw = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pmulhuw->addArg(x.value);
pmulhuw->addArg(y.value);
::basicBlock->appendInst(pmulhuw);
return RValue<UShort4>(V(result));
}
}
RValue<Int4> MulHigh(RValue<Int4> x, RValue<Int4> y)
{
// TODO: For x86, build an intrinsics version of this which uses shuffles + pmuludq.
// Scalarized implementation.
Int4 result;
result = Insert(result, Int((Long(Extract(x, 0)) * Long(Extract(y, 0))) >> Long(Int(32))), 0);
result = Insert(result, Int((Long(Extract(x, 1)) * Long(Extract(y, 1))) >> Long(Int(32))), 1);
result = Insert(result, Int((Long(Extract(x, 2)) * Long(Extract(y, 2))) >> Long(Int(32))), 2);
result = Insert(result, Int((Long(Extract(x, 3)) * Long(Extract(y, 3))) >> Long(Int(32))), 3);
return result;
}
RValue<UInt4> MulHigh(RValue<UInt4> x, RValue<UInt4> y)
{
// TODO: For x86, build an intrinsics version of this which uses shuffles + pmuludq.
if(false) // Partial product based implementation.
{
auto xh = x >> 16;
auto yh = y >> 16;
auto xl = x & UInt4(0x0000FFFF);
auto yl = y & UInt4(0x0000FFFF);
auto xlyh = xl * yh;
auto xhyl = xh * yl;
auto xlyhh = xlyh >> 16;
auto xhylh = xhyl >> 16;
auto xlyhl = xlyh & UInt4(0x0000FFFF);
auto xhyll = xhyl & UInt4(0x0000FFFF);
auto xlylh = (xl * yl) >> 16;
auto oflow = (xlyhl + xhyll + xlylh) >> 16;
return (xh * yh) + (xlyhh + xhylh) + oflow;
}
// Scalarized implementation.
Int4 result;
result = Insert(result, Int((Long(UInt(Extract(As<Int4>(x), 0))) * Long(UInt(Extract(As<Int4>(y), 0)))) >> Long(Int(32))), 0);
result = Insert(result, Int((Long(UInt(Extract(As<Int4>(x), 1))) * Long(UInt(Extract(As<Int4>(y), 1)))) >> Long(Int(32))), 1);
result = Insert(result, Int((Long(UInt(Extract(As<Int4>(x), 2))) * Long(UInt(Extract(As<Int4>(y), 2)))) >> Long(Int(32))), 2);
result = Insert(result, Int((Long(UInt(Extract(As<Int4>(x), 3))) * Long(UInt(Extract(As<Int4>(y), 3)))) >> Long(Int(32))), 3);
return As<UInt4>(result);
}
RValue<UShort4> Average(RValue<UShort4> x, RValue<UShort4> y)
{
assert(false && "UNIMPLEMENTED"); return RValue<UShort4>(V(nullptr));
}
Type *UShort4::getType()
{
return T(Type_v4i16);
}
RValue<Short> Extract(RValue<Short8> val, int i)
{
return RValue<Short>(Nucleus::createExtractElement(val.value, Short::getType(), i));
}
RValue<Short8> Insert(RValue<Short8> val, RValue<Short> element, int i)
{
return RValue<Short8>(Nucleus::createInsertElement(val.value, element.value, i));
}
RValue<Short8> operator<<(RValue<Short8> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Short8 result;
result = Insert(result, Extract(lhs, 0) << Short(rhs), 0);
result = Insert(result, Extract(lhs, 1) << Short(rhs), 1);
result = Insert(result, Extract(lhs, 2) << Short(rhs), 2);
result = Insert(result, Extract(lhs, 3) << Short(rhs), 3);
result = Insert(result, Extract(lhs, 4) << Short(rhs), 4);
result = Insert(result, Extract(lhs, 5) << Short(rhs), 5);
result = Insert(result, Extract(lhs, 6) << Short(rhs), 6);
result = Insert(result, Extract(lhs, 7) << Short(rhs), 7);
return result;
}
else
{
return RValue<Short8>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Short8> operator>>(RValue<Short8> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Short8 result;
result = Insert(result, Extract(lhs, 0) >> Short(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> Short(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> Short(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> Short(rhs), 3);
result = Insert(result, Extract(lhs, 4) >> Short(rhs), 4);
result = Insert(result, Extract(lhs, 5) >> Short(rhs), 5);
result = Insert(result, Extract(lhs, 6) >> Short(rhs), 6);
result = Insert(result, Extract(lhs, 7) >> Short(rhs), 7);
return result;
}
else
{
return RValue<Short8>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Int4> MulAdd(RValue<Short8> x, RValue<Short8> y)
{
assert(false && "UNIMPLEMENTED"); return RValue<Int4>(V(nullptr));
}
RValue<Short8> MulHigh(RValue<Short8> x, RValue<Short8> y)
{
assert(false && "UNIMPLEMENTED"); return RValue<Short8>(V(nullptr));
}
Type *Short8::getType()
{
return T(Ice::IceType_v8i16);
}
RValue<UShort> Extract(RValue<UShort8> val, int i)
{
return RValue<UShort>(Nucleus::createExtractElement(val.value, UShort::getType(), i));
}
RValue<UShort8> Insert(RValue<UShort8> val, RValue<UShort> element, int i)
{
return RValue<UShort8>(Nucleus::createInsertElement(val.value, element.value, i));
}
RValue<UShort8> operator<<(RValue<UShort8> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UShort8 result;
result = Insert(result, Extract(lhs, 0) << UShort(rhs), 0);
result = Insert(result, Extract(lhs, 1) << UShort(rhs), 1);
result = Insert(result, Extract(lhs, 2) << UShort(rhs), 2);
result = Insert(result, Extract(lhs, 3) << UShort(rhs), 3);
result = Insert(result, Extract(lhs, 4) << UShort(rhs), 4);
result = Insert(result, Extract(lhs, 5) << UShort(rhs), 5);
result = Insert(result, Extract(lhs, 6) << UShort(rhs), 6);
result = Insert(result, Extract(lhs, 7) << UShort(rhs), 7);
return result;
}
else
{
return RValue<UShort8>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UShort8> operator>>(RValue<UShort8> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UShort8 result;
result = Insert(result, Extract(lhs, 0) >> UShort(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> UShort(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> UShort(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> UShort(rhs), 3);
result = Insert(result, Extract(lhs, 4) >> UShort(rhs), 4);
result = Insert(result, Extract(lhs, 5) >> UShort(rhs), 5);
result = Insert(result, Extract(lhs, 6) >> UShort(rhs), 6);
result = Insert(result, Extract(lhs, 7) >> UShort(rhs), 7);
return result;
}
else
{
return RValue<UShort8>(Nucleus::createLShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UShort8> Swizzle(RValue<UShort8> x, char select0, char select1, char select2, char select3, char select4, char select5, char select6, char select7)
{
assert(false && "UNIMPLEMENTED"); return RValue<UShort8>(V(nullptr));
}
RValue<UShort8> MulHigh(RValue<UShort8> x, RValue<UShort8> y)
{
assert(false && "UNIMPLEMENTED"); return RValue<UShort8>(V(nullptr));
}
// FIXME: Implement as Shuffle(x, y, Select(i0, ..., i16)) and Shuffle(x, y, SELECT_PACK_REPEAT(element))
// RValue<UShort8> PackRepeat(RValue<Byte16> x, RValue<Byte16> y, int element)
// {
// assert(false && "UNIMPLEMENTED"); return RValue<UShort8>(V(nullptr));
// }
Type *UShort8::getType()
{
return T(Ice::IceType_v8i16);
}
RValue<Int> operator++(Int &val, int) // Post-increment
{
RValue<Int> res = val;
val += 1;
return res;
}
const Int &operator++(Int &val) // Pre-increment
{
val += 1;
return val;
}
RValue<Int> operator--(Int &val, int) // Post-decrement
{
RValue<Int> res = val;
val -= 1;
return res;
}
const Int &operator--(Int &val) // Pre-decrement
{
val -= 1;
return val;
}
RValue<Int> RoundInt(RValue<Float> cast)
{
if(emulateIntrinsics || CPUID::ARM)
{
// Push the fractional part off the mantissa. Accurate up to +/-2^22.
return Int((cast + Float(0x00C00000)) - Float(0x00C00000));
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Nearbyint, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto nearbyint = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
nearbyint->addArg(cast.value);
::basicBlock->appendInst(nearbyint);
return RValue<Int>(V(result));
}
}
Type *Int::getType()
{
return T(Ice::IceType_i32);
}
Type *Long::getType()
{
return T(Ice::IceType_i64);
}
UInt::UInt(RValue<Float> cast)
{
// Smallest positive value representable in UInt, but not in Int
const unsigned int ustart = 0x80000000u;
const float ustartf = float(ustart);
// If the value is negative, store 0, otherwise store the result of the conversion
storeValue((~(As<Int>(cast) >> 31) &
// Check if the value can be represented as an Int
IfThenElse(cast >= ustartf,
// If the value is too large, subtract ustart and re-add it after conversion.
As<Int>(As<UInt>(Int(cast - Float(ustartf))) + UInt(ustart)),
// Otherwise, just convert normally
Int(cast))).value);
}
RValue<UInt> operator++(UInt &val, int) // Post-increment
{
RValue<UInt> res = val;
val += 1;
return res;
}
const UInt &operator++(UInt &val) // Pre-increment
{
val += 1;
return val;
}
RValue<UInt> operator--(UInt &val, int) // Post-decrement
{
RValue<UInt> res = val;
val -= 1;
return res;
}
const UInt &operator--(UInt &val) // Pre-decrement
{
val -= 1;
return val;
}
// RValue<UInt> RoundUInt(RValue<Float> cast)
// {
// assert(false && "UNIMPLEMENTED"); return RValue<UInt>(V(nullptr));
// }
Type *UInt::getType()
{
return T(Ice::IceType_i32);
}
// Int2::Int2(RValue<Int> cast)
// {
// Value *extend = Nucleus::createZExt(cast.value, Long::getType());
// Value *vector = Nucleus::createBitCast(extend, Int2::getType());
//
// Constant *shuffle[2];
// shuffle[0] = Nucleus::createConstantInt(0);
// shuffle[1] = Nucleus::createConstantInt(0);
//
// Value *replicate = Nucleus::createShuffleVector(vector, UndefValue::get(Int2::getType()), Nucleus::createConstantVector(shuffle, 2));
//
// storeValue(replicate);
// }
RValue<Int2> operator<<(RValue<Int2> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Int2 result;
result = Insert(result, Extract(lhs, 0) << Int(rhs), 0);
result = Insert(result, Extract(lhs, 1) << Int(rhs), 1);
return result;
}
else
{
return RValue<Int2>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Int2> operator>>(RValue<Int2> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Int2 result;
result = Insert(result, Extract(lhs, 0) >> Int(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> Int(rhs), 1);
return result;
}
else
{
return RValue<Int2>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
Type *Int2::getType()
{
return T(Type_v2i32);
}
RValue<UInt> Extract(RValue<UInt2> val, int i)
{
return RValue<UInt>(Nucleus::createExtractElement(val.value, UInt::getType(), i));
}
RValue<UInt2> Insert(RValue<UInt2> val, RValue<UInt> element, int i)
{
return RValue<UInt2>(Nucleus::createInsertElement(val.value, element.value, i));
}
RValue<UInt2> operator<<(RValue<UInt2> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UInt2 result;
result = Insert(result, Extract(lhs, 0) << UInt(rhs), 0);
result = Insert(result, Extract(lhs, 1) << UInt(rhs), 1);
return result;
}
else
{
return RValue<UInt2>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UInt2> operator>>(RValue<UInt2> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UInt2 result;
result = Insert(result, Extract(lhs, 0) >> UInt(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> UInt(rhs), 1);
return result;
}
else
{
return RValue<UInt2>(Nucleus::createLShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
Type *UInt2::getType()
{
return T(Type_v2i32);
}
Int4::Int4(RValue<Byte4> cast) : XYZW(this)
{
Value *x = Nucleus::createBitCast(cast.value, Int::getType());
Value *a = Nucleus::createInsertElement(loadValue(), x, 0);
Value *e;
int swizzle[16] = {0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23};
Value *b = Nucleus::createBitCast(a, Byte16::getType());
Value *c = Nucleus::createShuffleVector(b, V(Nucleus::createNullValue(Byte16::getType())), swizzle);
int swizzle2[8] = {0, 8, 1, 9, 2, 10, 3, 11};
Value *d = Nucleus::createBitCast(c, Short8::getType());
e = Nucleus::createShuffleVector(d, V(Nucleus::createNullValue(Short8::getType())), swizzle2);
Value *f = Nucleus::createBitCast(e, Int4::getType());
storeValue(f);
}
Int4::Int4(RValue<SByte4> cast) : XYZW(this)
{
Value *x = Nucleus::createBitCast(cast.value, Int::getType());
Value *a = Nucleus::createInsertElement(loadValue(), x, 0);
int swizzle[16] = {0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6, 7, 7};
Value *b = Nucleus::createBitCast(a, Byte16::getType());
Value *c = Nucleus::createShuffleVector(b, b, swizzle);
int swizzle2[8] = {0, 0, 1, 1, 2, 2, 3, 3};
Value *d = Nucleus::createBitCast(c, Short8::getType());
Value *e = Nucleus::createShuffleVector(d, d, swizzle2);
*this = As<Int4>(e) >> 24;
}
Int4::Int4(RValue<Short4> cast) : XYZW(this)
{
int swizzle[8] = {0, 0, 1, 1, 2, 2, 3, 3};
Value *c = Nucleus::createShuffleVector(cast.value, cast.value, swizzle);
*this = As<Int4>(c) >> 16;
}
Int4::Int4(RValue<UShort4> cast) : XYZW(this)
{
int swizzle[8] = {0, 8, 1, 9, 2, 10, 3, 11};
Value *c = Nucleus::createShuffleVector(cast.value, Short8(0, 0, 0, 0, 0, 0, 0, 0).loadValue(), swizzle);
Value *d = Nucleus::createBitCast(c, Int4::getType());
storeValue(d);
}
Int4::Int4(RValue<Int> rhs) : XYZW(this)
{
Value *vector = Nucleus::createBitCast(rhs.value, Int4::getType());
int swizzle[4] = {0, 0, 0, 0};
Value *replicate = Nucleus::createShuffleVector(vector, vector, swizzle);
storeValue(replicate);
}
RValue<Int4> operator<<(RValue<Int4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Int4 result;
result = Insert(result, Extract(lhs, 0) << Int(rhs), 0);
result = Insert(result, Extract(lhs, 1) << Int(rhs), 1);
result = Insert(result, Extract(lhs, 2) << Int(rhs), 2);
result = Insert(result, Extract(lhs, 3) << Int(rhs), 3);
return result;
}
else
{
return RValue<Int4>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Int4> operator>>(RValue<Int4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
Int4 result;
result = Insert(result, Extract(lhs, 0) >> Int(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> Int(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> Int(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> Int(rhs), 3);
return result;
}
else
{
return RValue<Int4>(Nucleus::createAShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<Int4> CmpEQ(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpEQ(x.value, y.value));
}
RValue<Int4> CmpLT(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpSLT(x.value, y.value));
}
RValue<Int4> CmpLE(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpSLE(x.value, y.value));
}
RValue<Int4> CmpNEQ(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpNE(x.value, y.value));
}
RValue<Int4> CmpNLT(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpSGE(x.value, y.value));
}
RValue<Int4> CmpNLE(RValue<Int4> x, RValue<Int4> y)
{
return RValue<Int4>(Nucleus::createICmpSGT(x.value, y.value));
}
RValue<Int4> Max(RValue<Int4> x, RValue<Int4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Sle, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4i32);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<Int4>(V(result));
}
RValue<Int4> Min(RValue<Int4> x, RValue<Int4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Sgt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4i32);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<Int4>(V(result));
}
RValue<Int4> RoundInt(RValue<Float4> cast)
{
if(emulateIntrinsics || CPUID::ARM)
{
// Push the fractional part off the mantissa. Accurate up to +/-2^22.
return Int4((cast + Float4(0x00C00000)) - Float4(0x00C00000));
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Nearbyint, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto nearbyint = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
nearbyint->addArg(cast.value);
::basicBlock->appendInst(nearbyint);
return RValue<Int4>(V(result));
}
}
RValue<Short8> PackSigned(RValue<Int4> x, RValue<Int4> y)
{
if(emulateIntrinsics)
{
Short8 result;
result = Insert(result, SaturateSigned(Extract(x, 0)), 0);
result = Insert(result, SaturateSigned(Extract(x, 1)), 1);
result = Insert(result, SaturateSigned(Extract(x, 2)), 2);
result = Insert(result, SaturateSigned(Extract(x, 3)), 3);
result = Insert(result, SaturateSigned(Extract(y, 0)), 4);
result = Insert(result, SaturateSigned(Extract(y, 1)), 5);
result = Insert(result, SaturateSigned(Extract(y, 2)), 6);
result = Insert(result, SaturateSigned(Extract(y, 3)), 7);
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::VectorPackSigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pack = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pack->addArg(x.value);
pack->addArg(y.value);
::basicBlock->appendInst(pack);
return RValue<Short8>(V(result));
}
}
RValue<UShort8> PackUnsigned(RValue<Int4> x, RValue<Int4> y)
{
if(emulateIntrinsics || !(CPUID::SSE4_1 || CPUID::ARM))
{
RValue<Int4> sx = As<Int4>(x);
RValue<Int4> bx = (sx & ~(sx >> 31)) - Int4(0x8000);
RValue<Int4> sy = As<Int4>(y);
RValue<Int4> by = (sy & ~(sy >> 31)) - Int4(0x8000);
return As<UShort8>(PackSigned(bx, by) + Short8(0x8000u));
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v8i16);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::VectorPackUnsigned, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto pack = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
pack->addArg(x.value);
pack->addArg(y.value);
::basicBlock->appendInst(pack);
return RValue<UShort8>(V(result));
}
}
RValue<Int> SignMask(RValue<Int4> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
Int4 xx = (x >> 31) & Int4(0x00000001, 0x00000002, 0x00000004, 0x00000008);
return Extract(xx, 0) | Extract(xx, 1) | Extract(xx, 2) | Extract(xx, 3);
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SignMask, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto movmsk = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
movmsk->addArg(x.value);
::basicBlock->appendInst(movmsk);
return RValue<Int>(V(result));
}
}
Type *Int4::getType()
{
return T(Ice::IceType_v4i32);
}
UInt4::UInt4(RValue<Float4> cast) : XYZW(this)
{
// Smallest positive value representable in UInt, but not in Int
const unsigned int ustart = 0x80000000u;
const float ustartf = float(ustart);
// Check if the value can be represented as an Int
Int4 uiValue = CmpNLT(cast, Float4(ustartf));
// If the value is too large, subtract ustart and re-add it after conversion.
uiValue = (uiValue & As<Int4>(As<UInt4>(Int4(cast - Float4(ustartf))) + UInt4(ustart))) |
// Otherwise, just convert normally
(~uiValue & Int4(cast));
// If the value is negative, store 0, otherwise store the result of the conversion
storeValue((~(As<Int4>(cast) >> 31) & uiValue).value);
}
RValue<UInt4> operator<<(RValue<UInt4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UInt4 result;
result = Insert(result, Extract(lhs, 0) << UInt(rhs), 0);
result = Insert(result, Extract(lhs, 1) << UInt(rhs), 1);
result = Insert(result, Extract(lhs, 2) << UInt(rhs), 2);
result = Insert(result, Extract(lhs, 3) << UInt(rhs), 3);
return result;
}
else
{
return RValue<UInt4>(Nucleus::createShl(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UInt4> operator>>(RValue<UInt4> lhs, unsigned char rhs)
{
if(emulateIntrinsics)
{
UInt4 result;
result = Insert(result, Extract(lhs, 0) >> UInt(rhs), 0);
result = Insert(result, Extract(lhs, 1) >> UInt(rhs), 1);
result = Insert(result, Extract(lhs, 2) >> UInt(rhs), 2);
result = Insert(result, Extract(lhs, 3) >> UInt(rhs), 3);
return result;
}
else
{
return RValue<UInt4>(Nucleus::createLShr(lhs.value, V(::context->getConstantInt32(rhs))));
}
}
RValue<UInt4> CmpEQ(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpEQ(x.value, y.value));
}
RValue<UInt4> CmpLT(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpULT(x.value, y.value));
}
RValue<UInt4> CmpLE(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpULE(x.value, y.value));
}
RValue<UInt4> CmpNEQ(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpNE(x.value, y.value));
}
RValue<UInt4> CmpNLT(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpUGE(x.value, y.value));
}
RValue<UInt4> CmpNLE(RValue<UInt4> x, RValue<UInt4> y)
{
return RValue<UInt4>(Nucleus::createICmpUGT(x.value, y.value));
}
RValue<UInt4> Max(RValue<UInt4> x, RValue<UInt4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Ule, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4i32);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<UInt4>(V(result));
}
RValue<UInt4> Min(RValue<UInt4> x, RValue<UInt4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstIcmp::create(::function, Ice::InstIcmp::Ugt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4i32);
auto select = Ice::InstSelect::create(::function, result, condition, y.value, x.value);
::basicBlock->appendInst(select);
return RValue<UInt4>(V(result));
}
Type *UInt4::getType()
{
return T(Ice::IceType_v4i32);
}
Type *Half::getType()
{
return T(Ice::IceType_i16);
}
RValue<Float> Rcp_pp(RValue<Float> x, bool exactAtPow2)
{
return 1.0f / x;
}
RValue<Float> RcpSqrt_pp(RValue<Float> x)
{
return Rcp_pp(Sqrt(x));
}
RValue<Float> Sqrt(RValue<Float> x)
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Sqrt, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto sqrt = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
sqrt->addArg(x.value);
::basicBlock->appendInst(sqrt);
return RValue<Float>(V(result));
}
RValue<Float> Round(RValue<Float> x)
{
return Float4(Round(Float4(x))).x;
}
RValue<Float> Trunc(RValue<Float> x)
{
return Float4(Trunc(Float4(x))).x;
}
RValue<Float> Frac(RValue<Float> x)
{
return Float4(Frac(Float4(x))).x;
}
RValue<Float> Floor(RValue<Float> x)
{
return Float4(Floor(Float4(x))).x;
}
RValue<Float> Ceil(RValue<Float> x)
{
return Float4(Ceil(Float4(x))).x;
}
Type *Float::getType()
{
return T(Ice::IceType_f32);
}
Type *Float2::getType()
{
return T(Type_v2f32);
}
Float4::Float4(RValue<Float> rhs) : XYZW(this)
{
Value *vector = Nucleus::createBitCast(rhs.value, Float4::getType());
int swizzle[4] = {0, 0, 0, 0};
Value *replicate = Nucleus::createShuffleVector(vector, vector, swizzle);
storeValue(replicate);
}
RValue<Float4> Max(RValue<Float4> x, RValue<Float4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstFcmp::create(::function, Ice::InstFcmp::Ogt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
auto select = Ice::InstSelect::create(::function, result, condition, x.value, y.value);
::basicBlock->appendInst(select);
return RValue<Float4>(V(result));
}
RValue<Float4> Min(RValue<Float4> x, RValue<Float4> y)
{
Ice::Variable *condition = ::function->makeVariable(Ice::IceType_v4i1);
auto cmp = Ice::InstFcmp::create(::function, Ice::InstFcmp::Olt, condition, x.value, y.value);
::basicBlock->appendInst(cmp);
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
auto select = Ice::InstSelect::create(::function, result, condition, x.value, y.value);
::basicBlock->appendInst(select);
return RValue<Float4>(V(result));
}
RValue<Float4> Rcp_pp(RValue<Float4> x, bool exactAtPow2)
{
return Float4(1.0f) / x;
}
RValue<Float4> RcpSqrt_pp(RValue<Float4> x)
{
return Rcp_pp(Sqrt(x));
}
RValue<Float4> Sqrt(RValue<Float4> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
Float4 result;
result.x = Sqrt(Float(Float4(x).x));
result.y = Sqrt(Float(Float4(x).y));
result.z = Sqrt(Float(Float4(x).z));
result.w = Sqrt(Float(Float4(x).w));
return result;
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Sqrt, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto sqrt = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
sqrt->addArg(x.value);
::basicBlock->appendInst(sqrt);
return RValue<Float4>(V(result));
}
}
RValue<Int> SignMask(RValue<Float4> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
Int4 xx = (As<Int4>(x) >> 31) & Int4(0x00000001, 0x00000002, 0x00000004, 0x00000008);
return Extract(xx, 0) | Extract(xx, 1) | Extract(xx, 2) | Extract(xx, 3);
}
else
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_i32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::SignMask, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto movmsk = Ice::InstIntrinsicCall::create(::function, 1, result, target, intrinsic);
movmsk->addArg(x.value);
::basicBlock->appendInst(movmsk);
return RValue<Int>(V(result));
}
}
RValue<Int4> CmpEQ(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpOEQ(x.value, y.value));
}
RValue<Int4> CmpLT(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpOLT(x.value, y.value));
}
RValue<Int4> CmpLE(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpOLE(x.value, y.value));
}
RValue<Int4> CmpNEQ(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpONE(x.value, y.value));
}
RValue<Int4> CmpNLT(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpOGE(x.value, y.value));
}
RValue<Int4> CmpNLE(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpOGT(x.value, y.value));
}
RValue<Int4> CmpUEQ(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpUEQ(x.value, y.value));
}
RValue<Int4> CmpULT(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpULT(x.value, y.value));
}
RValue<Int4> CmpULE(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpULE(x.value, y.value));
}
RValue<Int4> CmpUNEQ(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpUNE(x.value, y.value));
}
RValue<Int4> CmpUNLT(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpUGE(x.value, y.value));
}
RValue<Int4> CmpUNLE(RValue<Float4> x, RValue<Float4> y)
{
return RValue<Int4>(Nucleus::createFCmpUGT(x.value, y.value));
}
RValue<Float4> Round(RValue<Float4> x)
{
if(emulateIntrinsics || CPUID::ARM)
{
// Push the fractional part off the mantissa. Accurate up to +/-2^22.
return (x + Float4(0x00C00000)) - Float4(0x00C00000);
}
else if(CPUID::SSE4_1)
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Round, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto round = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
round->addArg(x.value);
round->addArg(::context->getConstantInt32(0));
::basicBlock->appendInst(round);
return RValue<Float4>(V(result));
}
else
{
return Float4(RoundInt(x));
}
}
RValue<Float4> Trunc(RValue<Float4> x)
{
if(CPUID::SSE4_1)
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Round, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto round = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
round->addArg(x.value);
round->addArg(::context->getConstantInt32(3));
::basicBlock->appendInst(round);
return RValue<Float4>(V(result));
}
else
{
return Float4(Int4(x));
}
}
RValue<Float4> Frac(RValue<Float4> x)
{
Float4 frc;
if(CPUID::SSE4_1)
{
frc = x - Floor(x);
}
else
{
frc = x - Float4(Int4(x)); // Signed fractional part.
frc += As<Float4>(As<Int4>(CmpNLE(Float4(0.0f), frc)) & As<Int4>(Float4(1, 1, 1, 1))); // Add 1.0 if negative.
}
// x - floor(x) can be 1.0 for very small negative x.
// Clamp against the value just below 1.0.
return Min(frc, As<Float4>(Int4(0x3F7FFFFF)));
}
RValue<Float4> Floor(RValue<Float4> x)
{
if(CPUID::SSE4_1)
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Round, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto round = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
round->addArg(x.value);
round->addArg(::context->getConstantInt32(1));
::basicBlock->appendInst(round);
return RValue<Float4>(V(result));
}
else
{
return x - Frac(x);
}
}
RValue<Float4> Ceil(RValue<Float4> x)
{
if(CPUID::SSE4_1)
{
Ice::Variable *result = ::function->makeVariable(Ice::IceType_v4f32);
const Ice::Intrinsics::IntrinsicInfo intrinsic = {Ice::Intrinsics::Round, Ice::Intrinsics::SideEffects_F, Ice::Intrinsics::ReturnsTwice_F, Ice::Intrinsics::MemoryWrite_F};
auto target = ::context->getConstantUndef(Ice::IceType_i32);
auto round = Ice::InstIntrinsicCall::create(::function, 2, result, target, intrinsic);
round->addArg(x.value);
round->addArg(::context->getConstantInt32(2));
::basicBlock->appendInst(round);
return RValue<Float4>(V(result));
}
else
{
return -Floor(-x);
}
}
Type *Float4::getType()
{
return T(Ice::IceType_v4f32);
}
RValue<Long> Ticks()
{
assert(false && "UNIMPLEMENTED"); return RValue<Long>(V(nullptr));
}
// Below are functions currently unimplemented for the Subzero backend.
// They are stubbed to satisfy the linker.
#ifdef UNIMPLEMENTED
#undef UNIMPLEMENTED
#endif
#define UNIMPLEMENTED(msg) assert(((void)(msg), false))
RValue<Float4> Sin(RValue<Float4> x) { UNIMPLEMENTED("Subzero Sin()"); return Float4(0); }
RValue<Float4> Cos(RValue<Float4> x) { UNIMPLEMENTED("Subzero Cos()"); return Float4(0); }
RValue<Float4> Tan(RValue<Float4> x) { UNIMPLEMENTED("Subzero Tan()"); return Float4(0); }
RValue<Float4> Asin(RValue<Float4> x) { UNIMPLEMENTED("Subzero Asin()"); return Float4(0); }
RValue<Float4> Acos(RValue<Float4> x) { UNIMPLEMENTED("Subzero Acos()"); return Float4(0); }
RValue<Float4> Atan(RValue<Float4> x) { UNIMPLEMENTED("Subzero Atan()"); return Float4(0); }
RValue<Float4> Sinh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Sinh()"); return Float4(0); }
RValue<Float4> Cosh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Cosh()"); return Float4(0); }
RValue<Float4> Tanh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Tanh()"); return Float4(0); }
RValue<Float4> Asinh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Asinh()"); return Float4(0); }
RValue<Float4> Acosh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Acosh()"); return Float4(0); }
RValue<Float4> Atanh(RValue<Float4> x) { UNIMPLEMENTED("Subzero Atanh()"); return Float4(0); }
RValue<Float4> Atan2(RValue<Float4> x, RValue<Float4> y) { UNIMPLEMENTED("Subzero Atan2()"); return Float4(0); }
RValue<Float4> Pow(RValue<Float4> x, RValue<Float4> y) { UNIMPLEMENTED("Subzero Pow()"); return Float4(0); }
RValue<Float4> Exp(RValue<Float4> x) { UNIMPLEMENTED("Subzero Exp()"); return Float4(0); }
RValue<Float4> Log(RValue<Float4> x) { UNIMPLEMENTED("Subzero Log()"); return Float4(0); }
RValue<Float4> Exp2(RValue<Float4> x) { UNIMPLEMENTED("Subzero Exp2()"); return Float4(0); }
RValue<Float4> Log2(RValue<Float4> x) { UNIMPLEMENTED("Subzero Log2()"); return Float4(0); }
RValue<UInt4> Ctlz(RValue<UInt4> x, bool isZeroUndef) { UNIMPLEMENTED("Subzero Ctlz()"); return UInt4(0); }
RValue<UInt4> Cttz(RValue<UInt4> x, bool isZeroUndef) { UNIMPLEMENTED("Subzero Cttz()"); return UInt4(0); }
}