2659 lines
78 KiB
C++
2659 lines
78 KiB
C++
/**
|
|
******************************************************************************
|
|
* Xenia : Xbox 360 Emulator Research Project *
|
|
******************************************************************************
|
|
* Copyright 2013 Ben Vanik. All rights reserved. *
|
|
* Released under the BSD license - see LICENSE in the root for more details. *
|
|
******************************************************************************
|
|
*/
|
|
|
|
#include <alloy/backend/x64/lowering/lowering_sequences.h>
|
|
|
|
#include <alloy/backend/x64/x64_emitter.h>
|
|
#include <alloy/backend/x64/x64_function.h>
|
|
#include <alloy/backend/x64/lowering/lowering_table.h>
|
|
#include <alloy/runtime/symbol_info.h>
|
|
#include <alloy/runtime/runtime.h>
|
|
#include <alloy/runtime/thread_state.h>
|
|
|
|
using namespace alloy;
|
|
using namespace alloy::backend::x64;
|
|
using namespace alloy::backend::x64::lowering;
|
|
using namespace alloy::hir;
|
|
using namespace alloy::runtime;
|
|
|
|
using namespace Xbyak;
|
|
|
|
namespace {
|
|
|
|
#define UNIMPLEMENTED_SEQ() __debugbreak()
|
|
#define ASSERT_INVALID_TYPE() XEASSERTALWAYS()
|
|
|
|
#define ITRACE 1
|
|
#define DTRACE 0
|
|
|
|
#define SHUFPS_SWAP_DWORDS 0x1B
|
|
|
|
// A note about vectors:
|
|
// Alloy represents vectors as xyzw pairs, with indices 0123.
|
|
// XMM registers are xyzw pairs with indices 3210, making them more like wzyx.
|
|
// This makes things somewhat confusing. It'd be nice to just shuffle the
|
|
// registers around on load/store, however certain operations require that
|
|
// data be in the right offset.
|
|
// Basically, this identity must hold:
|
|
// shuffle(vec, b00011011) -> {x,y,z,w} => {x,y,z,w}
|
|
// All indices and operations must respect that.
|
|
|
|
// TODO(benvanik): emit traces/printfs/etc
|
|
|
|
void Dummy() {
|
|
//
|
|
}
|
|
|
|
void PrintString(void* raw_context, const char* str) {
|
|
// TODO(benvanik): generate this thunk at runtime? or a shim?
|
|
auto thread_state = *((ThreadState**)raw_context);
|
|
fprintf(stdout, "XE[t] :%d: %s\n", thread_state->GetThreadID(), str);
|
|
fflush(stdout);
|
|
}
|
|
|
|
void TraceContextLoad(void* raw_context, uint64_t offset, uint64_t value) {
|
|
fprintf(stdout, "%lld (%.llX) = ctx i64 +%lld\n", (int64_t)value, value, offset);
|
|
fflush(stdout);
|
|
}
|
|
void TraceContextStore(void* raw_context, uint64_t offset, uint64_t value) {
|
|
fprintf(stdout, "ctx i64 +%lld = %lld (%.llX)\n", offset, (int64_t)value, value);
|
|
fflush(stdout);
|
|
}
|
|
|
|
uint64_t LoadClock(void* raw_context) {
|
|
LARGE_INTEGER counter;
|
|
uint64_t time = 0;
|
|
if (QueryPerformanceCounter(&counter)) {
|
|
time = counter.QuadPart;
|
|
}
|
|
return time;
|
|
}
|
|
|
|
void CallNative(X64Emitter& e, void* target) {
|
|
e.mov(e.rax, (uint64_t)target);
|
|
e.call(e.rax);
|
|
e.mov(e.rcx, e.qword[e.rsp + 0]);
|
|
e.mov(e.rdx, e.qword[e.rcx + 8]); // membase
|
|
}
|
|
|
|
// TODO(benvanik): fancy stuff.
|
|
void* ResolveFunctionSymbol(void* raw_context, FunctionInfo* symbol_info) {
|
|
// TODO(benvanik): generate this thunk at runtime? or a shim?
|
|
auto thread_state = *((ThreadState**)raw_context);
|
|
|
|
Function* fn = NULL;
|
|
thread_state->runtime()->ResolveFunction(symbol_info->address(), &fn);
|
|
XEASSERTNOTNULL(fn);
|
|
XEASSERT(fn->type() == Function::USER_FUNCTION);
|
|
auto x64_fn = (X64Function*)fn;
|
|
return x64_fn->machine_code();
|
|
}
|
|
void* ResolveFunctionAddress(void* raw_context, uint64_t target_address) {
|
|
// TODO(benvanik): generate this thunk at runtime? or a shim?
|
|
auto thread_state = *((ThreadState**)raw_context);
|
|
|
|
Function* fn = NULL;
|
|
thread_state->runtime()->ResolveFunction(target_address, &fn);
|
|
XEASSERTNOTNULL(fn);
|
|
XEASSERT(fn->type() == Function::USER_FUNCTION);
|
|
auto x64_fn = (X64Function*)fn;
|
|
return x64_fn->machine_code();
|
|
}
|
|
void IssueCall(X64Emitter& e, FunctionInfo* symbol_info, uint32_t flags) {
|
|
// If we are an extern function, we can directly insert a call.
|
|
auto fn = symbol_info->function();
|
|
if (fn && fn->type() == Function::EXTERN_FUNCTION) {
|
|
auto extern_fn = (ExternFunction*)fn;
|
|
e.mov(e.rdx, (uint64_t)extern_fn->arg0());
|
|
e.mov(e.r8, (uint64_t)extern_fn->arg1());
|
|
e.mov(e.rax, (uint64_t)extern_fn->handler());
|
|
} else {
|
|
// Generic call, resolve address.
|
|
// TODO(benvanik): caching/etc. For now this makes debugging easier.
|
|
e.mov(e.rdx, (uint64_t)symbol_info);
|
|
e.mov(e.rax, (uint64_t)ResolveFunctionSymbol);
|
|
e.call(e.rax);
|
|
e.mov(e.rcx, e.qword[e.rsp + 0]);
|
|
e.mov(e.rdx, e.qword[e.rcx + 8]); // membase
|
|
}
|
|
if (flags & CALL_TAIL) {
|
|
// TODO(benvanik): adjust stack?
|
|
e.add(e.rsp, 0x40);
|
|
e.jmp(e.rax);
|
|
} else {
|
|
e.call(e.rax);
|
|
e.mov(e.rcx, e.qword[e.rsp + 0]);
|
|
e.mov(e.rdx, e.qword[e.rcx + 8]); // membase
|
|
}
|
|
}
|
|
void IssueCallIndirect(X64Emitter& e, Value* target, uint32_t flags) {
|
|
Reg64 r;
|
|
e.BeginOp(target, r, 0);
|
|
if (r != e.rdx) {
|
|
e.mov(e.rdx, r);
|
|
}
|
|
e.EndOp(r);
|
|
e.mov(e.rax, (uint64_t)ResolveFunctionAddress);
|
|
e.call(e.rax);
|
|
e.mov(e.rcx, e.qword[e.rsp + 0]);
|
|
e.mov(e.rdx, e.qword[e.rcx + 8]); // membase
|
|
if (flags & CALL_TAIL) {
|
|
// TODO(benvanik): adjust stack?
|
|
e.add(e.rsp, 0x40);
|
|
e.jmp(e.rax);
|
|
} else {
|
|
e.call(e.rax);
|
|
e.mov(e.rcx, e.qword[e.rsp + 0]);
|
|
e.mov(e.rdx, e.qword[e.rcx + 8]); // membase
|
|
}
|
|
}
|
|
|
|
// Sets EFLAGs with zf for the given value.
|
|
// ZF = 1 if false, 0 = true (so jz = jump if false)
|
|
void CheckBoolean(X64Emitter& e, Value* v) {
|
|
if (v->IsConstant()) {
|
|
e.mov(e.ah, (v->IsConstantZero() ? 1 : 0) << 6);
|
|
e.sahf();
|
|
} else if (v->type == INT8_TYPE) {
|
|
Reg8 src;
|
|
e.BeginOp(v, src, 0);
|
|
e.test(src, src);
|
|
e.EndOp(src);
|
|
} else if (v->type == INT16_TYPE) {
|
|
Reg16 src;
|
|
e.BeginOp(v, src, 0);
|
|
e.test(src, src);
|
|
e.EndOp(src);
|
|
} else if (v->type == INT32_TYPE) {
|
|
Reg32 src;
|
|
e.BeginOp(v, src, 0);
|
|
e.test(src, src);
|
|
e.EndOp(src);
|
|
} else if (v->type == INT64_TYPE) {
|
|
Reg64 src;
|
|
e.BeginOp(v, src, 0);
|
|
e.test(src, src);
|
|
e.EndOp(src);
|
|
} else if (v->type == FLOAT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (v->type == FLOAT64_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (v->type == VEC128_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
}
|
|
|
|
void CompareXX(X64Emitter& e, Instr*& i, void(set_fn)(X64Emitter& e, Reg8& dest, bool invert)) {
|
|
if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8, SIG_TYPE_I8)) {
|
|
Reg8 dest;
|
|
Reg8 src1, src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src1, src2);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8, SIG_TYPE_I8C)) {
|
|
Reg8 dest;
|
|
Reg8 src1;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
e.cmp(src1, i->src2.value->constant.i8);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8C, SIG_TYPE_I8)) {
|
|
Reg8 dest;
|
|
Reg8 src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src2, i->src1.value->constant.i8);
|
|
set_fn(e, dest, true);
|
|
e.EndOp(dest, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16, SIG_TYPE_I16)) {
|
|
Reg8 dest;
|
|
Reg16 src1, src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src1, src2);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16, SIG_TYPE_I16C)) {
|
|
Reg8 dest;
|
|
Reg16 src1;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
e.cmp(src1, i->src2.value->constant.i16);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16C, SIG_TYPE_I16)) {
|
|
Reg8 dest;
|
|
Reg16 src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src2, i->src1.value->constant.i16);
|
|
e.sete(dest);
|
|
set_fn(e, dest, true);
|
|
e.EndOp(dest, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32, SIG_TYPE_I32)) {
|
|
Reg8 dest;
|
|
Reg32 src1, src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src1, src2);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32, SIG_TYPE_I32C)) {
|
|
Reg8 dest;
|
|
Reg32 src1;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
e.cmp(src1, i->src2.value->constant.i32);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32C, SIG_TYPE_I32)) {
|
|
Reg8 dest;
|
|
Reg32 src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src2, i->src1.value->constant.i32);
|
|
set_fn(e, dest, true);
|
|
e.EndOp(dest, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64, SIG_TYPE_I64)) {
|
|
Reg8 dest;
|
|
Reg64 src1, src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
e.cmp(src1, src2);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64, SIG_TYPE_I64C)) {
|
|
Reg8 dest;
|
|
Reg64 src1;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
e.mov(e.rax, i->src2.value->constant.i64);
|
|
e.cmp(src1, e.rax);
|
|
set_fn(e, dest, false);
|
|
e.EndOp(dest, src1);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64C, SIG_TYPE_I64)) {
|
|
Reg8 dest;
|
|
Reg64 src2;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src2.value, src2, 0);
|
|
e.mov(e.rax, i->src1.value->constant.i64);
|
|
e.cmp(src2, e.rax);
|
|
set_fn(e, dest, true);
|
|
e.EndOp(dest, src2);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
};
|
|
|
|
typedef void(v_fn)(X64Emitter& e, Instr& i, const Reg& dest_src);
|
|
template<typename T>
|
|
void IntUnaryOpV(X64Emitter& e, Instr*& i, v_fn v_fn,
|
|
T& dest, T& src1) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
if (dest == src1) {
|
|
v_fn(e, *i, dest);
|
|
} else {
|
|
e.mov(dest, src1);
|
|
v_fn(e, *i, dest);
|
|
}
|
|
e.EndOp(dest, src1);
|
|
}
|
|
template<typename CT, typename T>
|
|
void IntUnaryOpC(X64Emitter& e, Instr*& i, v_fn v_fn,
|
|
T& dest, Value* src1) {
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, (uint64_t)src1->get_constant(CT()));
|
|
v_fn(e, *i, dest);
|
|
e.EndOp(dest);
|
|
}
|
|
void IntUnaryOp(X64Emitter& e, Instr*& i, v_fn v_fn) {
|
|
if (i->Match(SIG_TYPE_I8, SIG_TYPE_I8)) {
|
|
Reg8 dest, src1;
|
|
IntUnaryOpV(e, i, v_fn, dest, src1);
|
|
} else if (i->Match(SIG_TYPE_I8, SIG_TYPE_I8C)) {
|
|
Reg8 dest;
|
|
IntUnaryOpC<int8_t>(e, i, v_fn, dest, i->src1.value);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16)) {
|
|
Reg16 dest, src1;
|
|
IntUnaryOpV(e, i, v_fn, dest, src1);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16C)) {
|
|
Reg16 dest;
|
|
IntUnaryOpC<int16_t>(e, i, v_fn, dest, i->src1.value);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32)) {
|
|
Reg32 dest, src1;
|
|
IntUnaryOpV(e, i, v_fn, dest, src1);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32C)) {
|
|
Reg32 dest;
|
|
IntUnaryOpC<int32_t>(e, i, v_fn, dest, i->src1.value);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64)) {
|
|
Reg64 dest, src1;
|
|
IntUnaryOpV(e, i, v_fn, dest, src1);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64C)) {
|
|
Reg64 dest;
|
|
IntUnaryOpC<int64_t>(e, i, v_fn, dest, i->src1.value);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
if (i->flags & ARITHMETIC_SET_CARRY) {
|
|
// EFLAGS should have CA set?
|
|
// (so long as we don't fuck with it)
|
|
// UNIMPLEMENTED_SEQ();
|
|
}
|
|
};
|
|
|
|
typedef void(vv_fn)(X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src);
|
|
typedef void(vc_fn)(X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src);
|
|
template<typename TD, typename TS1, typename TS2>
|
|
void IntBinaryOpVV(X64Emitter& e, Instr*& i, vv_fn vv_fn,
|
|
TD& dest, TS1& src1, TS2& src2) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
if (dest == src1) {
|
|
vv_fn(e, *i, dest, src2);
|
|
} else if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
vv_fn(e, *i, dest, src1);
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src1);
|
|
vv_fn(e, *i, e.rax, src2);
|
|
e.mov(dest, e.rax);
|
|
}
|
|
} else {
|
|
e.mov(dest, src1);
|
|
vv_fn(e, *i, dest, src2);
|
|
}
|
|
e.EndOp(dest, src1, src2);
|
|
}
|
|
template<typename CT, typename TD, typename TS1>
|
|
void IntBinaryOpVC(X64Emitter& e, Instr*& i, vv_fn vv_fn, vc_fn vc_fn,
|
|
TD& dest, TS1& src1, Value* src2) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
if (dest.getBit() <= 32) {
|
|
// 32-bit.
|
|
if (dest == src1) {
|
|
vc_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()));
|
|
} else {
|
|
e.mov(dest, src1);
|
|
vc_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()));
|
|
}
|
|
} else {
|
|
// 64-bit.
|
|
if (dest == src1) {
|
|
e.mov(e.rax, src2->constant.i64);
|
|
vv_fn(e, *i, dest, e.rax);
|
|
} else {
|
|
e.mov(e.rax, src2->constant.i64);
|
|
e.mov(dest, src1);
|
|
vv_fn(e, *i, dest, e.rax);
|
|
}
|
|
}
|
|
e.EndOp(dest, src1);
|
|
}
|
|
template<typename CT, typename TD, typename TS2>
|
|
void IntBinaryOpCV(X64Emitter& e, Instr*& i, vv_fn vv_fn, vc_fn vc_fn,
|
|
TD& dest, Value* src1, TS2& src2) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src2.value, src2, 0);
|
|
if (dest.getBit() <= 32) {
|
|
// 32-bit.
|
|
if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
vc_fn(e, *i, dest, (uint32_t)src1->get_constant(CT()));
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src2);
|
|
e.mov(dest, (uint32_t)src1->get_constant(CT()));
|
|
vv_fn(e, *i, dest, e.rax);
|
|
}
|
|
} else {
|
|
e.mov(dest, src2);
|
|
vc_fn(e, *i, dest, (uint32_t)src1->get_constant(CT()));
|
|
}
|
|
} else {
|
|
// 64-bit.
|
|
if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
e.mov(e.rax, src1->constant.i64);
|
|
vv_fn(e, *i, dest, e.rax);
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src1->constant.i64);
|
|
vv_fn(e, *i, e.rax, src2);
|
|
e.mov(dest, e.rax);
|
|
}
|
|
} else {
|
|
e.mov(e.rax, src2);
|
|
e.mov(dest, src1->constant.i64);
|
|
vv_fn(e, *i, dest, e.rax);
|
|
}
|
|
}
|
|
e.EndOp(dest, src2);
|
|
}
|
|
void IntBinaryOp(X64Emitter& e, Instr*& i, vv_fn vv_fn, vc_fn vc_fn) {
|
|
// TODO(benvanik): table lookup. This linear scan is slow.
|
|
// Note: we assume DEST.type = SRC1.type, but that SRC2.type may vary.
|
|
XEASSERT(i->dest->type == i->src1.value->type);
|
|
if (i->Match(SIG_TYPE_I8, SIG_TYPE_I8, SIG_TYPE_I8)) {
|
|
Reg8 dest, src1, src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I8, SIG_TYPE_I8, SIG_TYPE_I8C)) {
|
|
Reg8 dest, src1;
|
|
IntBinaryOpVC<int8_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I8, SIG_TYPE_I8C, SIG_TYPE_I8)) {
|
|
Reg8 dest, src2;
|
|
IntBinaryOpCV<int8_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I16)) {
|
|
Reg16 dest, src1, src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I16C)) {
|
|
Reg16 dest, src1;
|
|
IntBinaryOpVC<int16_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16C, SIG_TYPE_I16)) {
|
|
Reg16 dest, src2;
|
|
IntBinaryOpCV<int16_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I32)) {
|
|
Reg32 dest, src1, src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I32C)) {
|
|
Reg32 dest, src1;
|
|
IntBinaryOpVC<int32_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32C, SIG_TYPE_I32)) {
|
|
Reg32 dest, src2;
|
|
IntBinaryOpCV<int32_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I64)) {
|
|
Reg64 dest, src1, src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I64C)) {
|
|
Reg64 dest, src1;
|
|
IntBinaryOpVC<int64_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64C, SIG_TYPE_I64)) {
|
|
Reg64 dest, src2;
|
|
IntBinaryOpCV<int64_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
// Start forced src2=i8
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I8)) {
|
|
Reg16 dest, src1;
|
|
Reg8 src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I8C)) {
|
|
Reg16 dest, src1;
|
|
IntBinaryOpVC<int8_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16C, SIG_TYPE_I8)) {
|
|
Reg16 dest;
|
|
Reg8 src2;
|
|
IntBinaryOpCV<int16_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I8)) {
|
|
Reg32 dest, src1;
|
|
Reg8 src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I8C)) {
|
|
Reg32 dest, src1;
|
|
IntBinaryOpVC<int8_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32C, SIG_TYPE_I8)) {
|
|
Reg32 dest;
|
|
Reg8 src2;
|
|
IntBinaryOpCV<int32_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I8)) {
|
|
Reg64 dest, src1;
|
|
Reg8 src2;
|
|
IntBinaryOpVV(e, i, vv_fn, dest, src1, src2);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I8C)) {
|
|
Reg64 dest, src1;
|
|
IntBinaryOpVC<int8_t>(e, i, vv_fn, vc_fn, dest, src1, i->src2.value);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64C, SIG_TYPE_I8)) {
|
|
Reg64 dest;
|
|
Reg8 src2;
|
|
IntBinaryOpCV<int64_t>(e, i, vv_fn, vc_fn, dest, i->src1.value, src2);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
if (i->flags & ARITHMETIC_SET_CARRY) {
|
|
// EFLAGS should have CA set?
|
|
// (so long as we don't fuck with it)
|
|
// UNIMPLEMENTED_SEQ();
|
|
}
|
|
};
|
|
|
|
typedef void(vvv_fn)(X64Emitter& e, Instr& i, const Reg& dest_src1, const Operand& src2, const Operand& src3);
|
|
typedef void(vvc_fn)(X64Emitter& e, Instr& i, const Reg& dest_src1, const Operand& src2, uint32_t src3);
|
|
typedef void(vcv_fn)(X64Emitter& e, Instr& i, const Reg& dest_src1, uint32_t src2, const Operand& src3);
|
|
template<typename TD, typename TS1, typename TS2, typename TS3>
|
|
void IntTernaryOpVVV(X64Emitter& e, Instr*& i, vvv_fn vvv_fn,
|
|
TD& dest, TS1& src1, TS2& src2, TS3& src3) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0,
|
|
i->src3.value, src3, 0);
|
|
if (dest == src1) {
|
|
vvv_fn(e, *i, dest, src2, src3);
|
|
} else if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
vvv_fn(e, *i, dest, src1, src3);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
} else {
|
|
e.mov(dest, src1);
|
|
vvv_fn(e, *i, dest, src2, src3);
|
|
}
|
|
e.EndOp(dest, src1, src2, src3);
|
|
}
|
|
template<typename CT, typename TD, typename TS1, typename TS2>
|
|
void IntTernaryOpVVC(X64Emitter& e, Instr*& i, vvv_fn vvv_fn, vvc_fn vvc_fn,
|
|
TD& dest, TS1& src1, TS2& src2, Value* src3) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src2.value, src2, 0);
|
|
if (dest.getBit() <= 32) {
|
|
// 32-bit.
|
|
if (dest == src1) {
|
|
vvc_fn(e, *i, dest, src2, (uint32_t)src3->get_constant(CT()));
|
|
} else if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
vvc_fn(e, *i, dest, src1, (uint32_t)src3->get_constant(CT()));
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src2);
|
|
e.mov(dest, src1);
|
|
vvc_fn(e, *i, dest, e.rax, (uint32_t)src3->get_constant(CT()));
|
|
}
|
|
} else {
|
|
e.mov(dest, src1);
|
|
vvc_fn(e, *i, dest, src2, (uint32_t)src3->get_constant(CT()));
|
|
}
|
|
} else {
|
|
// 64-bit.
|
|
if (dest == src1) {
|
|
e.mov(e.rax, src3->constant.i64);
|
|
vvv_fn(e, *i, dest, src2, e.rax);
|
|
} else if (dest == src2) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
e.mov(e.rax, src3->constant.i64);
|
|
vvv_fn(e, *i, dest, src1, e.rax);
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src1);
|
|
e.mov(src1, src2);
|
|
e.mov(dest, e.rax);
|
|
e.mov(e.rax, src3->constant.i64);
|
|
vvv_fn(e, *i, dest, src1, e.rax);
|
|
}
|
|
} else {
|
|
e.mov(e.rax, src3->constant.i64);
|
|
e.mov(dest, src1);
|
|
vvv_fn(e, *i, dest, src2, e.rax);
|
|
}
|
|
}
|
|
e.EndOp(dest, src1, src2);
|
|
}
|
|
template<typename CT, typename TD, typename TS1, typename TS3>
|
|
void IntTernaryOpVCV(X64Emitter& e, Instr*& i, vvv_fn vvv_fn, vcv_fn vcv_fn,
|
|
TD& dest, TS1& src1, Value* src2, TS3& src3) {
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0,
|
|
i->src3.value, src3, 0);
|
|
if (dest.getBit() <= 32) {
|
|
// 32-bit.
|
|
if (dest == src1) {
|
|
vcv_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()), src3);
|
|
} else if (dest == src3) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
vcv_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()), src1);
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src3);
|
|
e.mov(dest, src1);
|
|
vcv_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()), e.rax);
|
|
}
|
|
} else {
|
|
e.mov(dest, src1);
|
|
vcv_fn(e, *i, dest, (uint32_t)src2->get_constant(CT()), src3);
|
|
}
|
|
} else {
|
|
// 64-bit.
|
|
if (dest == src1) {
|
|
e.mov(e.rax, src2->constant.i64);
|
|
vvv_fn(e, *i, dest, e.rax, src3);
|
|
} else if (dest == src3) {
|
|
if (i->opcode->flags & OPCODE_FLAG_COMMUNATIVE) {
|
|
e.mov(e.rax, src2->constant.i64);
|
|
vvv_fn(e, *i, dest, src1, e.rax);
|
|
} else {
|
|
// Eww.
|
|
e.mov(e.rax, src1);
|
|
e.mov(src1, src3);
|
|
e.mov(dest, e.rax);
|
|
e.mov(e.rax, src2->constant.i64);
|
|
vvv_fn(e, *i, dest, e.rax, src1);
|
|
}
|
|
} else {
|
|
e.mov(e.rax, src2->constant.i64);
|
|
e.mov(dest, src1);
|
|
vvv_fn(e, *i, dest, e.rax, src3);
|
|
}
|
|
}
|
|
e.EndOp(dest, src1, src3);
|
|
}
|
|
void IntTernaryOp(X64Emitter& e, Instr*& i, vvv_fn vvv_fn, vvc_fn vvc_fn, vcv_fn vcv_fn) {
|
|
// TODO(benvanik): table lookup. This linear scan is slow.
|
|
// Note: we assume DEST.type = SRC1.type = SRC2.type, but that SRC3.type may vary.
|
|
XEASSERT(i->dest->type == i->src1.value->type &&
|
|
i->dest->type == i->src2.value->type);
|
|
// TODO(benvanik): table lookup.
|
|
if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8, SIG_TYPE_I8, SIG_TYPE_I8)) {
|
|
Reg8 dest, src1, src2;
|
|
Reg8 src3;
|
|
IntTernaryOpVVV(e, i, vvv_fn, dest, src1, src2, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8, SIG_TYPE_I8, SIG_TYPE_I8C)) {
|
|
Reg8 dest, src1, src2;
|
|
IntTernaryOpVVC<int8_t>(e, i, vvv_fn, vvc_fn, dest, src1, src2, i->src3.value);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I8)) {
|
|
Reg16 dest, src1, src2;
|
|
Reg8 src3;
|
|
IntTernaryOpVVV(e, i, vvv_fn, dest, src1, src2, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16, SIG_TYPE_I16, SIG_TYPE_I8C)) {
|
|
Reg16 dest, src1, src2;
|
|
IntTernaryOpVVC<int8_t>(e, i, vvv_fn, vvc_fn, dest, src1, src2, i->src3.value);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I8)) {
|
|
Reg32 dest, src1, src2;
|
|
Reg8 src3;
|
|
IntTernaryOpVVV(e, i,vvv_fn, dest, src1, src2, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32, SIG_TYPE_I32, SIG_TYPE_I8C)) {
|
|
Reg32 dest, src1, src2;
|
|
IntTernaryOpVVC<int8_t>(e, i, vvv_fn, vvc_fn, dest, src1, src2, i->src3.value);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I8)) {
|
|
Reg64 dest, src1, src2;
|
|
Reg8 src3;
|
|
IntTernaryOpVVV(e, i, vvv_fn, dest, src1, src2, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64, SIG_TYPE_I64, SIG_TYPE_I8C)) {
|
|
Reg64 dest, src1, src2;
|
|
IntTernaryOpVVC<int8_t>(e, i, vvv_fn, vvc_fn, dest, src1, src2, i->src3.value);
|
|
//
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8, SIG_TYPE_I8C, SIG_TYPE_I8)) {
|
|
Reg8 dest, src1, src3;
|
|
IntTernaryOpVCV<int8_t>(e, i, vvv_fn, vcv_fn, dest, src1, i->src2.value, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16, SIG_TYPE_I16C, SIG_TYPE_I8)) {
|
|
Reg16 dest, src1, src3;
|
|
IntTernaryOpVCV<int16_t>(e, i, vvv_fn, vcv_fn, dest, src1, i->src2.value, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32, SIG_TYPE_I32C, SIG_TYPE_I8)) {
|
|
Reg32 dest, src1, src3;
|
|
IntTernaryOpVCV<int32_t>(e, i, vvv_fn, vcv_fn, dest, src1, i->src2.value, src3);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64, SIG_TYPE_I64C, SIG_TYPE_I8)) {
|
|
Reg64 dest, src1, src3;
|
|
IntTernaryOpVCV<int64_t>(e, i, vvv_fn, vcv_fn, dest, src1, i->src2.value, src3);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
if (i->flags & ARITHMETIC_SET_CARRY) {
|
|
// EFLAGS should have CA set?
|
|
// (so long as we don't fuck with it)
|
|
// UNIMPLEMENTED_SEQ();
|
|
}
|
|
}
|
|
|
|
} // namespace
|
|
|
|
|
|
void alloy::backend::x64::lowering::RegisterSequences(LoweringTable* table) {
|
|
// --------------------------------------------------------------------------
|
|
// General
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_COMMENT, [](X64Emitter& e, Instr*& i) {
|
|
#if ITRACE
|
|
// TODO(benvanik): pass through.
|
|
// TODO(benvanik): don't just leak this memory.
|
|
auto str = (const char*)i->src1.offset;
|
|
auto str_copy = xestrdupa(str);
|
|
e.mov(e.rdx, (uint64_t)str_copy);
|
|
CallNative(e, PrintString);
|
|
#endif // ITRACE
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_NOP, [](X64Emitter& e, Instr*& i) {
|
|
// If we got this, chances are we want it.
|
|
e.nop();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Debugging
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_SOURCE_OFFSET, [](X64Emitter& e, Instr*& i) {
|
|
#if XE_DEBUG
|
|
e.nop();
|
|
e.nop();
|
|
e.mov(e.eax, (uint32_t)i->src1.offset);
|
|
e.nop();
|
|
e.nop();
|
|
#endif // XE_DEBUG
|
|
|
|
e.MarkSourceOffset(i);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DEBUG_BREAK, [](X64Emitter& e, Instr*& i) {
|
|
// TODO(benvanik): insert a call to the debug break function to let the
|
|
// debugger know.
|
|
e.db(0xCC);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DEBUG_BREAK_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
e.inLocalLabel();
|
|
CheckBoolean(e, i->src1.value);
|
|
e.jz(".x", e.T_SHORT);
|
|
// TODO(benvanik): insert a call to the debug break function to let the
|
|
// debugger know.
|
|
e.db(0xCC);
|
|
e.L(".x");
|
|
e.outLocalLabel();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_TRAP, [](X64Emitter& e, Instr*& i) {
|
|
// TODO(benvanik): insert a call to the trap function to let the
|
|
// debugger know.
|
|
e.db(0xCC);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_TRAP_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
e.inLocalLabel();
|
|
CheckBoolean(e, i->src1.value);
|
|
e.jz(".x", e.T_SHORT);
|
|
// TODO(benvanik): insert a call to the trap function to let the
|
|
// debugger know.
|
|
e.db(0xCC);
|
|
e.L(".x");
|
|
e.outLocalLabel();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Calls
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_CALL, [](X64Emitter& e, Instr*& i) {
|
|
IssueCall(e, i->src1.symbol_info, i->flags);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CALL_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
e.inLocalLabel();
|
|
CheckBoolean(e, i->src1.value);
|
|
e.jz(".x", e.T_SHORT);
|
|
IssueCall(e, i->src2.symbol_info, i->flags);
|
|
e.L(".x");
|
|
e.outLocalLabel();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CALL_INDIRECT, [](X64Emitter& e, Instr*& i) {
|
|
IssueCallIndirect(e, i->src1.value, i->flags);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CALL_INDIRECT_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
e.inLocalLabel();
|
|
CheckBoolean(e, i->src1.value);
|
|
e.jz(".x", e.T_SHORT);
|
|
IssueCallIndirect(e, i->src2.value, i->flags);
|
|
e.L(".x");
|
|
e.outLocalLabel();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_RETURN, [](X64Emitter& e, Instr*& i) {
|
|
// If this is the last instruction in the last block, just let us
|
|
// fall through.
|
|
if (i->next || i->block->next) {
|
|
e.jmp("epilog", CodeGenerator::T_NEAR);
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_RETURN_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
CheckBoolean(e, i->src1.value);
|
|
e.jnz("epilog", CodeGenerator::T_NEAR);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Branches
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_BRANCH, [](X64Emitter& e, Instr*& i) {
|
|
auto target = i->src1.label;
|
|
e.jmp(target->name, e.T_NEAR);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_BRANCH_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
CheckBoolean(e, i->src1.value);
|
|
auto target = i->src2.label;
|
|
e.jnz(target->name, e.T_NEAR);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_BRANCH_FALSE, [](X64Emitter& e, Instr*& i) {
|
|
CheckBoolean(e, i->src1.value);
|
|
auto target = i->src2.label;
|
|
e.jz(target->name, e.T_NEAR);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Types
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_ASSIGN, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntUnaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src) {
|
|
// nop - the mov will have happened.
|
|
});
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CAST, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ZERO_EXTEND, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_I16, SIG_TYPE_I8)) {
|
|
Reg16 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movzx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I8)) {
|
|
Reg32 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movzx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I16)) {
|
|
Reg32 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movzx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I8)) {
|
|
Reg64 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movzx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I16)) {
|
|
Reg64 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movzx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I32)) {
|
|
Reg64 dest;
|
|
Reg32 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest.cvt32(), src.cvt32());
|
|
e.EndOp(dest, src);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SIGN_EXTEND, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_I16, SIG_TYPE_I8)) {
|
|
Reg16 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I8)) {
|
|
Reg32 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I16)) {
|
|
Reg32 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I8)) {
|
|
Reg64 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I16)) {
|
|
Reg64 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsx(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I32)) {
|
|
Reg64 dest;
|
|
Reg32 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.movsxd(dest, src);
|
|
e.EndOp(dest, src);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_TRUNCATE, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_I8, SIG_TYPE_I16)) {
|
|
Reg8 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt8());
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I8, SIG_TYPE_I32)) {
|
|
Reg8 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt8());
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I8, SIG_TYPE_I64)) {
|
|
Reg8 dest;
|
|
Reg64 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt8());
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I32)) {
|
|
Reg16 dest;
|
|
Reg32 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt16());
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_I64)) {
|
|
Reg16 dest;
|
|
Reg64 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt16());
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I64)) {
|
|
Reg32 dest;
|
|
Reg64 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.mov(dest, src.cvt32());
|
|
e.EndOp(dest, src);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CONVERT, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ROUND, [](X64Emitter& e, Instr*& i) {
|
|
// flags = ROUND_TO_*
|
|
if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_CONVERT_I2F, [](X64Emitter& e, Instr*& i) {
|
|
// flags = ARITHMETIC_SATURATE | ARITHMETIC_UNSIGNED
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_CONVERT_F2I, [](X64Emitter& e, Instr*& i) {
|
|
// flags = ARITHMETIC_SATURATE | ARITHMETIC_UNSIGNED
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Constants
|
|
// --------------------------------------------------------------------------
|
|
|
|
// specials for zeroing/etc (xor/etc)
|
|
|
|
table->AddSequence(OPCODE_LOAD_VECTOR_SHL, [](X64Emitter& e, Instr*& i) {
|
|
XEASSERT(i->dest->type == VEC128_TYPE);
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_LOAD_VECTOR_SHR, [](X64Emitter& e, Instr*& i) {
|
|
XEASSERT(i->dest->type == VEC128_TYPE);
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_LOAD_CLOCK, [](X64Emitter& e, Instr*& i) {
|
|
// It'd be cool to call QueryPerformanceCounter directly, but w/e.
|
|
CallNative(e, LoadClock);
|
|
Reg64 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.rax);
|
|
e.EndOp(dest);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Context
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_LOAD_CONTEXT, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_I8, SIG_TYPE_IGNORE)) {
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.byte[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8b, dest);
|
|
CallNative(e, TraceContextLoad);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_IGNORE)) {
|
|
Reg16 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.word[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8w, dest);
|
|
CallNative(e, TraceContextLoad);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_IGNORE)) {
|
|
Reg32 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.dword[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8d, dest);
|
|
CallNative(e, TraceContextLoad);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_IGNORE)) {
|
|
Reg64 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.qword[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8, dest);
|
|
CallNative(e, TraceContextLoad);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_F32, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.movss(dest, e.dword[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_F64, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.movsd(dest, e.qword[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_V128, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
// NOTE: we always know we are aligned.
|
|
e.movaps(dest, e.ptr[e.rcx + i->src1.offset]);
|
|
e.EndOp(dest);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_STORE_CONTEXT, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I8)) {
|
|
Reg8 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.byte[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8b, src);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I8C)) {
|
|
e.mov(e.byte[e.rcx + i->src1.offset], i->src2.value->constant.i8);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8b, i->src2.value->constant.i8);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I16)) {
|
|
Reg16 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.word[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8w, src);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I16C)) {
|
|
e.mov(e.word[e.rcx + i->src1.offset], i->src2.value->constant.i16);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8w, i->src2.value->constant.i16);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I32)) {
|
|
Reg32 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.dword[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8d, src);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I32C)) {
|
|
e.mov(e.dword[e.rcx + i->src1.offset], i->src2.value->constant.i32);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8d, i->src2.value->constant.i32);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I64)) {
|
|
Reg64 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.qword[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8, src);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I64C)) {
|
|
e.mov(e.qword[e.rcx + i->src1.offset], i->src2.value->constant.i64);
|
|
#if DTRACE
|
|
e.mov(e.rdx, i->src1.offset);
|
|
e.mov(e.r8, i->src2.value->constant.i64);
|
|
CallNative(e, TraceContextStore);
|
|
#endif // DTRACE
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F32)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.movss(e.dword[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F32C)) {
|
|
e.mov(e.dword[e.rcx + i->src1.offset], i->src2.value->constant.i32);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F64)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.movsd(e.qword[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F64C)) {
|
|
e.mov(e.qword[e.rcx + i->src1.offset], i->src2.value->constant.i64);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_V128)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
// NOTE: we always know we are aligned.
|
|
e.movaps(e.ptr[e.rcx + i->src1.offset], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_V128C)) {
|
|
e.mov(e.qword[e.rcx + i->src1.offset], i->src2.value->constant.v128.low);
|
|
e.mov(e.qword[e.rcx + i->src1.offset + 8], i->src2.value->constant.v128.high);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Memory
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_LOAD, [](X64Emitter& e, Instr*& i) {
|
|
// If this is a constant address load, check to see if it's in a register
|
|
// range. We'll also probably want a dynamic check for unverified loads.
|
|
// So far, most games use constants.
|
|
if (i->src1.value->IsConstant()) {
|
|
uint64_t address = i->src1.value->AsUint64();
|
|
auto cbs = e.runtime()->access_callbacks();
|
|
while (cbs) {
|
|
if (cbs->handles(cbs->context, address)) {
|
|
// Eh, hacking lambdas.
|
|
i->src3.offset = (uint64_t)cbs;
|
|
IntUnaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src) {
|
|
auto cbs = (RegisterAccessCallbacks*)i.src3.offset;
|
|
e.mov(e.rcx, (uint64_t)cbs->context);
|
|
e.mov(e.rdx, i.src1.value->AsUint64());
|
|
CallNative(e, cbs->read);
|
|
e.mov(dest_src, e.rax);
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
}
|
|
cbs = cbs->next;
|
|
}
|
|
}
|
|
|
|
// TODO(benvanik): dynamic register access check.
|
|
// mov reg, [membase + address.32]
|
|
Reg64 addr_off;
|
|
RegExp addr;
|
|
if (i->src1.value->IsConstant()) {
|
|
// TODO(benvanik): a way to do this without using a register.
|
|
e.mov(e.eax, i->src1.value->AsUint32());
|
|
addr = e.rdx + e.rax;
|
|
} else {
|
|
e.BeginOp(i->src1.value, addr_off, 0);
|
|
e.mov(addr_off.cvt32(), addr_off.cvt32()); // trunc to 32bits
|
|
addr = e.rdx + addr_off;
|
|
}
|
|
if (i->Match(SIG_TYPE_I8, SIG_TYPE_IGNORE)) {
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.byte[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_I16, SIG_TYPE_IGNORE)) {
|
|
Reg16 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.word[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_IGNORE)) {
|
|
Reg32 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.dword[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_IGNORE)) {
|
|
Reg64 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.mov(dest, e.qword[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_F32, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.movss(dest, e.dword[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_F64, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.movsd(dest, e.qword[addr]);
|
|
e.EndOp(dest);
|
|
} else if (i->Match(SIG_TYPE_V128, SIG_TYPE_IGNORE)) {
|
|
Xmm dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
// TODO(benvanik): we should try to stick to movaps if possible.
|
|
e.movups(dest, e.ptr[addr]);
|
|
e.EndOp(dest);
|
|
e.db(0xCC);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
if (!i->src1.value->IsConstant()) {
|
|
e.EndOp(addr_off);
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_STORE, [](X64Emitter& e, Instr*& i) {
|
|
// If this is a constant address store, check to see if it's in a
|
|
// register range. We'll also probably want a dynamic check for
|
|
// unverified stores. So far, most games use constants.
|
|
if (i->src1.value->IsConstant()) {
|
|
uint64_t address = i->src1.value->AsUint64();
|
|
auto cbs = e.runtime()->access_callbacks();
|
|
while (cbs) {
|
|
if (cbs->handles(cbs->context, address)) {
|
|
e.mov(e.rcx, (uint64_t)cbs->context);
|
|
e.mov(e.rdx, address);
|
|
if (i->src2.value->IsConstant()) {
|
|
e.mov(e.r8, i->src2.value->AsUint64());
|
|
} else {
|
|
Reg64 src2;
|
|
e.BeginOp(i->src2.value, src2, 0);
|
|
switch (i->src2.value->type) {
|
|
case INT8_TYPE:
|
|
e.movzx(e.r8d, src2.cvt8());
|
|
break;
|
|
case INT16_TYPE:
|
|
e.movzx(e.r8d, src2.cvt16());
|
|
break;
|
|
case INT32_TYPE:
|
|
e.movzx(e.r8, src2.cvt32());
|
|
break;
|
|
case INT64_TYPE:
|
|
e.mov(e.r8, src2);
|
|
break;
|
|
default: ASSERT_INVALID_TYPE(); break;
|
|
}
|
|
e.EndOp(src2);
|
|
}
|
|
// eh?
|
|
e.bswap(e.r8);
|
|
CallNative(e, cbs->write);
|
|
}
|
|
cbs = cbs->next;
|
|
}
|
|
}
|
|
|
|
// TODO(benvanik): dynamic register access check
|
|
// mov [membase + address.32], reg
|
|
Reg64 addr_off;
|
|
RegExp addr;
|
|
if (i->src1.value->IsConstant()) {
|
|
e.mov(e.eax, i->src1.value->AsUint32());
|
|
addr = e.rdx + e.rax;
|
|
} else {
|
|
e.BeginOp(i->src1.value, addr_off, 0);
|
|
e.mov(addr_off.cvt32(), addr_off.cvt32()); // trunc to 32bits
|
|
addr = e.rdx + addr_off;
|
|
}
|
|
if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I8)) {
|
|
Reg8 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.byte[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I8C)) {
|
|
e.mov(e.byte[addr], i->src2.value->constant.i8);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I16)) {
|
|
Reg16 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.word[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I16C)) {
|
|
e.mov(e.word[addr], i->src2.value->constant.i16);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I32)) {
|
|
Reg32 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.dword[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I32C)) {
|
|
e.mov(e.dword[addr], i->src2.value->constant.i32);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I64)) {
|
|
Reg64 src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.mov(e.qword[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_I64C)) {
|
|
e.mov(e.qword[addr], i->src2.value->constant.i64);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F32)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.movss(e.dword[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F32C)) {
|
|
e.mov(e.dword[addr], i->src2.value->constant.i32);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F64)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
e.movsd(e.qword[addr], src);
|
|
e.EndOp(src);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_F64C)) {
|
|
e.mov(e.qword[addr], i->src2.value->constant.i64);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_V128)) {
|
|
Xmm src;
|
|
e.BeginOp(i->src2.value, src, 0);
|
|
// TODO(benvanik): we should try to stick to movaps if possible.
|
|
e.movups(e.ptr[addr], src);
|
|
e.EndOp(src);
|
|
e.db(0xCC);
|
|
} else if (i->Match(SIG_TYPE_X, SIG_TYPE_IGNORE, SIG_TYPE_V128C)) {
|
|
e.mov(e.ptr[addr], i->src2.value->constant.v128.low);
|
|
e.mov(e.ptr[addr + 8], i->src2.value->constant.v128.high);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
if (!i->src1.value->IsConstant()) {
|
|
e.EndOp(addr_off);
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_PREFETCH, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Comparisons
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_MAX, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_MIN, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SELECT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_IS_TRUE, [](X64Emitter& e, Instr*& i) {
|
|
CheckBoolean(e, i->src1.value);
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.setnz(dest);
|
|
e.EndOp(dest);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_IS_FALSE, [](X64Emitter& e, Instr*& i) {
|
|
CheckBoolean(e, i->src1.value);
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.setz(dest);
|
|
e.EndOp(dest);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_EQ, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.sete(dest);
|
|
} else {
|
|
e.setne(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_NE, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setne(dest);
|
|
} else {
|
|
e.sete(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_SLT, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setl(dest);
|
|
} else {
|
|
e.setge(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_SLE, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setle(dest);
|
|
} else {
|
|
e.setg(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_SGT, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setg(dest);
|
|
} else {
|
|
e.setle(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_SGE, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setge(dest);
|
|
} else {
|
|
e.setl(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_ULT, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setb(dest);
|
|
} else {
|
|
e.setae(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_ULE, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setbe(dest);
|
|
} else {
|
|
e.seta(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_UGT, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.seta(dest);
|
|
} else {
|
|
e.setbe(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_COMPARE_UGE, [](X64Emitter& e, Instr*& i) {
|
|
CompareXX(e, i, [](X64Emitter& e, Reg8& dest, bool invert) {
|
|
if (!invert) {
|
|
e.setae(dest);
|
|
} else {
|
|
e.setb(dest);
|
|
}
|
|
});
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DID_CARRY, [](X64Emitter& e, Instr*& i) {
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.setc(dest);
|
|
e.EndOp(dest);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DID_OVERFLOW, [](X64Emitter& e, Instr*& i) {
|
|
Reg8 dest;
|
|
e.BeginOp(i->dest, dest, REG_DEST);
|
|
e.seto(dest);
|
|
e.EndOp(dest);
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DID_SATURATE, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_COMPARE_EQ, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_COMPARE_SGT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_COMPARE_SGE, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_COMPARE_UGT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_COMPARE_UGE, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Math
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_ADD, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
e.add(dest_src, src);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.add(dest_src, src);
|
|
});
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ADD_CARRY, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
// dest = src1 + src2 + src3.i8
|
|
IntTernaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src2, const Operand& src3) {
|
|
Reg8 src3_8(src3.getIdx());
|
|
if (src3.getIdx() <= 4) {
|
|
e.mov(e.ah, src3_8);
|
|
} else {
|
|
e.mov(e.al, src3_8);
|
|
e.mov(e.ah, e.al);
|
|
}
|
|
e.sahf();
|
|
e.adc(dest_src, src2);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src2, uint32_t src3) {
|
|
e.mov(e.eax, src3);
|
|
e.mov(e.ah, e.al);
|
|
e.sahf();
|
|
e.adc(dest_src, src2);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src2, const Operand& src3) {
|
|
Reg8 src3_8(src3.getIdx());
|
|
if (src3.getIdx() <= 4) {
|
|
e.mov(e.ah, src3_8);
|
|
} else {
|
|
e.mov(e.al, src3_8);
|
|
e.mov(e.ah, e.al);
|
|
}
|
|
e.sahf();
|
|
e.adc(dest_src, src2);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_ADD, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->flags == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == FLOAT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SUB, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
e.sub(dest_src, src);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.sub(dest_src, src);
|
|
});
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
#define LIKE_REG(dest, like) Operand(dest.getIdx(), dest.getKind(), like.getBit(), false)
|
|
|
|
table->AddSequence(OPCODE_MUL, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.mul(src);
|
|
} else {
|
|
e.imul(src);
|
|
}
|
|
e.mov(dest_src, Nax);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
auto Ndx = LIKE_REG(e.rdx, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
e.mov(Ndx, src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.mul(Ndx);
|
|
} else {
|
|
e.imul(Ndx);
|
|
}
|
|
e.mov(dest_src, Nax);
|
|
});
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_MUL_HI, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
auto Ndx = LIKE_REG(e.rdx, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.mul(src);
|
|
} else {
|
|
e.imul(src);
|
|
}
|
|
e.mov(dest_src, Ndx);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
auto Ndx = LIKE_REG(e.rdx, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
e.mov(Ndx, src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.mul(Ndx);
|
|
} else {
|
|
e.imul(Ndx);
|
|
}
|
|
e.mov(dest_src, Ndx);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DIV, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.div(src);
|
|
} else {
|
|
e.idiv(src);
|
|
}
|
|
e.mov(dest_src, Nax);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
// RAX = value, RDX = clobbered
|
|
// TODO(benvanik): make the register allocator put dest_src in RAX?
|
|
auto Nax = LIKE_REG(e.rax, dest_src);
|
|
auto Ndx = LIKE_REG(e.rdx, dest_src);
|
|
e.mov(Nax, dest_src);
|
|
e.mov(Ndx, src);
|
|
if (i.flags & ARITHMETIC_UNSIGNED) {
|
|
e.div(Ndx);
|
|
} else {
|
|
e.idiv(Ndx);
|
|
}
|
|
e.mov(dest_src, Nax);
|
|
});
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_MUL_ADD, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_MUL_SUB, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_NEG, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ABS, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SQRT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_RSQRT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_POW2, [](X64Emitter& e, Instr*& i) {
|
|
if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_LOG2, [](X64Emitter& e, Instr*& i) {
|
|
if (IsFloatType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DOT_PRODUCT_3, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_DOT_PRODUCT_4, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_AND, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
e.and(dest_src, src);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.and(dest_src, src);
|
|
});
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_OR, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
e.or(dest_src, src);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.or(dest_src, src);
|
|
});
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_XOR, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
e.xor(dest_src, src);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.xor(dest_src, src);
|
|
});
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_NOT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntUnaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src) {
|
|
e.not(dest_src);
|
|
});
|
|
} else if (IsVecType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SHL, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
// TODO(benvanik): use shlx if available.
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// Can only shl by cl. Eww x86.
|
|
Reg8 shamt(src.getIdx());
|
|
e.mov(e.rax, e.rcx);
|
|
e.mov(e.cl, shamt);
|
|
e.shl(dest_src, e.cl);
|
|
e.mov(e.rcx, e.rax);
|
|
// BeaEngine can't disasm this, boo.
|
|
/*Reg32e dest_src_e(dest_src.getIdx(), MAX(dest_src.getBit(), 32));
|
|
Reg32e src_e(src.getIdx(), MAX(dest_src.getBit(), 32));
|
|
e.and(src_e, 0x3F);
|
|
e.shlx(dest_src_e, dest_src_e, src_e);*/
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.shl(dest_src, src);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SHR, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
// TODO(benvanik): use shrx if available.
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// Can only sar by cl. Eww x86.
|
|
Reg8 shamt(src.getIdx());
|
|
e.mov(e.rax, e.rcx);
|
|
e.mov(e.cl, shamt);
|
|
e.shr(dest_src, e.cl);
|
|
e.mov(e.rcx, e.rax);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.shr(dest_src, src);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SHA, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
// TODO(benvanik): use sarx if available.
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// Can only sar by cl. Eww x86.
|
|
Reg8 shamt(src.getIdx());
|
|
e.mov(e.rax, e.rcx);
|
|
e.mov(e.cl, shamt);
|
|
e.sar(dest_src, e.cl);
|
|
e.mov(e.rcx, e.rax);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.sar(dest_src, src);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_SHL, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->flags == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_SHR, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->flags == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_VECTOR_SHA, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->flags == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ROTATE_LEFT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
IntBinaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, const Operand& src) {
|
|
// Can only rol by cl. Eww x86.
|
|
Reg8 shamt(src.getIdx());
|
|
e.mov(e.rax, e.rcx);
|
|
e.mov(e.cl, shamt);
|
|
e.rol(dest_src, e.cl);
|
|
e.mov(e.rcx, e.rax);
|
|
},
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src, uint32_t src) {
|
|
e.rol(dest_src, src);
|
|
});
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_BYTE_SWAP, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_I16, SIG_TYPE_I16)) {
|
|
Reg16 d, s1;
|
|
// TODO(benvanik): fix register allocator to put the value in ABCD
|
|
//e.BeginOp(i->dest, d, REG_DEST | REG_ABCD,
|
|
// i->src1.value, s1, 0);
|
|
//if (d != s1) {
|
|
// e.mov(d, s1);
|
|
// e.xchg(d.cvt8(), Reg8(d.getIdx() + 4));
|
|
//} else {
|
|
// e.xchg(d.cvt8(), Reg8(d.getIdx() + 4));
|
|
//}
|
|
e.BeginOp(i->dest, d, REG_DEST,
|
|
i->src1.value, s1, 0);
|
|
e.mov(e.ax, s1);
|
|
e.xchg(e.ah, e.al);
|
|
e.mov(d, e.ax);
|
|
e.EndOp(d, s1);
|
|
} else if (i->Match(SIG_TYPE_I32, SIG_TYPE_I32)) {
|
|
Reg32 d, s1;
|
|
e.BeginOp(i->dest, d, REG_DEST,
|
|
i->src1.value, s1, 0);
|
|
if (d != s1) {
|
|
e.mov(d, s1);
|
|
e.bswap(d);
|
|
} else {
|
|
e.bswap(d);
|
|
}
|
|
e.EndOp(d, s1);
|
|
} else if (i->Match(SIG_TYPE_I64, SIG_TYPE_I64)) {
|
|
Reg64 d, s1;
|
|
e.BeginOp(i->dest, d, REG_DEST,
|
|
i->src1.value, s1, 0);
|
|
if (d != s1) {
|
|
e.mov(d, s1);
|
|
e.bswap(d);
|
|
} else {
|
|
e.bswap(d);
|
|
}
|
|
e.EndOp(d, s1);
|
|
} else if (i->Match(SIG_TYPE_V128, SIG_TYPE_V128)) {
|
|
Xmm d, s1;
|
|
e.db(0xCC);
|
|
e.BeginOp(i->dest, d, REG_DEST,
|
|
i->src1.value, s1, 0);
|
|
if (d != s1) {
|
|
e.shufps(d, s1, SHUFPS_SWAP_DWORDS);
|
|
} else {
|
|
e.shufps(d, d, SHUFPS_SWAP_DWORDS);
|
|
}
|
|
e.EndOp(d, s1);
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_CNTLZ, [](X64Emitter& e, Instr*& i) {
|
|
if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I8)) {
|
|
Reg8 dest;
|
|
Reg8 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.bsr(dest.cvt16(), src.cvt16());
|
|
// ZF = 1 if zero
|
|
e.mov(e.eax, 16);
|
|
e.cmovz(dest.cvt32(), e.eax);
|
|
e.sub(dest, 8);
|
|
e.xor(dest, 0x7);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I16)) {
|
|
Reg8 dest;
|
|
Reg16 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.bsr(dest.cvt16(), src);
|
|
// ZF = 1 if zero
|
|
e.mov(e.eax, 16);
|
|
e.cmovz(dest.cvt32(), e.eax);
|
|
e.xor(dest, 0xF);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I32)) {
|
|
Reg8 dest;
|
|
Reg32 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.bsr(dest.cvt32(), src);
|
|
// ZF = 1 if zero
|
|
e.mov(e.eax, 32);
|
|
e.cmovz(dest.cvt32(), e.eax);
|
|
e.xor(dest, 0x1F);
|
|
e.EndOp(dest, src);
|
|
} else if (i->Match(SIG_TYPE_IGNORE, SIG_TYPE_I64)) {
|
|
Reg8 dest;
|
|
Reg64 src;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src, 0);
|
|
e.bsr(dest, src);
|
|
// ZF = 1 if zero
|
|
e.mov(e.eax, 64);
|
|
e.cmovz(dest.cvt32(), e.eax);
|
|
e.xor(dest, 0x3F);
|
|
e.EndOp(dest, src);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_INSERT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->src3.value->type == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src3.value->type == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src3.value->type == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_EXTRACT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->src1.value->type)) {
|
|
if (i->dest->type == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->dest->type == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->dest->type == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SPLAT, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->src1.value->type == INT8_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src1.value->type == INT16_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src1.value->type == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src1.value->type == FLOAT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_PERMUTE, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
if (i->src1.value->type == INT32_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->src1.value->type == VEC128_TYPE) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_SWIZZLE, [](X64Emitter& e, Instr*& i) {
|
|
if (IsVecType(i->dest->type)) {
|
|
// Defined by SWIZZLE_MASK()
|
|
if (i->flags == INT32_TYPE || i->flags == FLOAT32_TYPE) {
|
|
uint8_t swizzle_mask = (uint8_t)i->src2.offset;
|
|
e.db(0xCC);
|
|
Xmm dest, src1;
|
|
e.BeginOp(i->dest, dest, REG_DEST,
|
|
i->src1.value, src1, 0);
|
|
e.pshufd(dest, src1, swizzle_mask);
|
|
e.EndOp(dest, src1);
|
|
} else {
|
|
UNIMPLEMENTED_SEQ();
|
|
}
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_PACK, [](X64Emitter& e, Instr*& i) {
|
|
if (i->flags == PACK_TYPE_D3DCOLOR) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_FLOAT16_2) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_FLOAT16_4) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_SHORT_2) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S8_IN_16_LO) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S8_IN_16_HI) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S16_IN_32_LO) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S16_IN_32_HI) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_UNPACK, [](X64Emitter& e, Instr*& i) {
|
|
if (i->flags == PACK_TYPE_D3DCOLOR) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_FLOAT16_2) {
|
|
// 1 bit sign, 5 bit exponent, 10 bit mantissa
|
|
// D3D10 half float format
|
|
// TODO(benvanik): http://blogs.msdn.com/b/chuckw/archive/2012/09/11/directxmath-f16c-and-fma.aspx
|
|
// Use _mm_cvtph_ps -- requires very modern processors (SSE5+)
|
|
// Unpacking half floats: http://fgiesen.wordpress.com/2012/03/28/half-to-float-done-quic/
|
|
// Packing half floats: https://gist.github.com/rygorous/2156668
|
|
// Load source, move from tight pack of X16Y16.... to X16...Y16...
|
|
// Also zero out the high end.
|
|
// TODO(benvanik): special case constant unpacks that just get 0/1/etc.
|
|
IntUnaryOp(
|
|
e, i,
|
|
[](X64Emitter& e, Instr& i, const Reg& dest_src) {
|
|
// sx = src.iw >> 16;
|
|
// sy = src.iw & 0xFFFF;
|
|
// dest = { 3.0 + (sx / float(1 << 22)),
|
|
// 3.0 + (sy / float(1 << 22)),
|
|
// 0.0,
|
|
// 1.0); --- or 3.0?
|
|
// So:
|
|
// xmm = {0,0,0,packed}
|
|
// xmm <<= 1w {0,0,packed,0}
|
|
// xmm = VCVTPH2PS(xmm) {sx,sy,0,0}
|
|
// xmm /=
|
|
});
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_FLOAT16_4) {
|
|
// Could be shared with FLOAT16_2.
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_SHORT_2) {
|
|
// (VD.x) = 3.0 + (VB.x)*2^-22
|
|
// (VD.y) = 3.0 + (VB.y)*2^-22
|
|
// (VD.z) = 0.0
|
|
// (VD.w) = 3.0
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S8_IN_16_LO) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S8_IN_16_HI) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S16_IN_32_LO) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else if (i->flags == PACK_TYPE_S16_IN_32_HI) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
// --------------------------------------------------------------------------
|
|
// Atomic
|
|
// --------------------------------------------------------------------------
|
|
|
|
table->AddSequence(OPCODE_COMPARE_EXCHANGE, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ATOMIC_EXCHANGE, [](X64Emitter& e, Instr*& i) {
|
|
if (IsIntType(i->dest->type)) {
|
|
UNIMPLEMENTED_SEQ();
|
|
} else {
|
|
ASSERT_INVALID_TYPE();
|
|
}
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ATOMIC_ADD, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
|
|
table->AddSequence(OPCODE_ATOMIC_SUB, [](X64Emitter& e, Instr*& i) {
|
|
UNIMPLEMENTED_SEQ();
|
|
i = e.Advance(i);
|
|
return true;
|
|
});
|
|
}
|