rpcs3/Utilities/Thread.cpp

1424 lines
29 KiB
C++
Raw Normal View History

#include "stdafx.h"
2014-09-11 19:18:19 +00:00
#include "Log.h"
2015-01-16 14:36:53 +00:00
#include "rpcs3/Ini.h"
#include "Emu/System.h"
2015-03-04 21:51:14 +00:00
#include "Emu/CPU/CPUThreadManager.h"
#include "Emu/CPU/CPUThread.h"
2015-03-04 21:51:14 +00:00
#include "Emu/Cell/RawSPUThread.h"
2015-01-04 21:59:17 +00:00
#include "Emu/SysCalls/SysCalls.h"
#include "Thread.h"
#ifdef _WIN32
#include <windows.h>
#else
2015-02-13 21:45:36 +00:00
#ifdef __APPLE__
2015-02-11 04:17:39 +00:00
#define _XOPEN_SOURCE
#define __USE_GNU
2015-02-13 21:45:36 +00:00
#endif
#include <signal.h>
#include <ucontext.h>
#endif
void SetCurrentThreadDebugName(const char* threadName)
{
#if defined(_MSC_VER) // this is VS-specific way to set thread names for the debugger
#pragma pack(push,8)
struct THREADNAME_INFO
{
DWORD dwType;
LPCSTR szName;
DWORD dwThreadID;
DWORD dwFlags;
} info;
#pragma pack(pop)
info.dwType = 0x1000;
info.szName = threadName;
info.dwThreadID = -1;
info.dwFlags = 0;
__try
{
RaiseException(0x406D1388, 0, sizeof(info) / sizeof(ULONG_PTR), (ULONG_PTR*)&info);
}
__except (EXCEPTION_EXECUTE_HANDLER)
{
}
#endif
}
enum x64_reg_t : u32
{
2015-02-17 00:08:23 +00:00
X64R_RAX = 0,
2015-02-15 12:31:42 +00:00
X64R_RCX,
X64R_RDX,
X64R_RBX,
X64R_RSP,
X64R_RBP,
X64R_RSI,
X64R_RDI,
X64R_R8,
X64R_R9,
X64R_R10,
X64R_R11,
X64R_R12,
X64R_R13,
X64R_R14,
X64R_R15,
2015-02-17 00:08:23 +00:00
X64R_XMM0 = 0,
2015-02-15 12:31:42 +00:00
X64R_XMM1,
X64R_XMM2,
X64R_XMM3,
X64R_XMM4,
X64R_XMM5,
X64R_XMM6,
X64R_XMM7,
X64R_XMM8,
X64R_XMM9,
X64R_XMM10,
X64R_XMM11,
X64R_XMM12,
X64R_XMM13,
X64R_XMM14,
X64R_XMM15,
X64R_AL,
X64R_CL,
X64R_DL,
X64R_BL,
X64R_AH,
X64R_CH,
X64R_DH,
X64R_BH,
X64_NOT_SET,
X64_IMM8,
X64_IMM16,
X64_IMM32,
2015-02-15 12:31:42 +00:00
X64R_ECX = X64R_CL,
};
enum x64_op_t : u32
{
2015-02-15 17:13:06 +00:00
X64OP_NONE,
2015-03-04 21:51:14 +00:00
X64OP_LOAD, // obtain and put the value into x64 register
X64OP_STORE, // take the value from x64 register or an immediate and use it
2015-06-21 00:17:42 +00:00
// example: add eax,[rax] -> X64OP_LOAD_ADD (add the value to x64 register)
// example: add [rax],eax -> X64OP_LOAD_ADD_STORE (this will probably never happen for MMIO registers)
2015-02-15 12:31:42 +00:00
X64OP_MOVS,
2015-02-17 02:01:47 +00:00
X64OP_STOS,
2015-02-15 12:31:42 +00:00
X64OP_XCHG,
X64OP_CMPXCHG,
2015-06-21 00:17:42 +00:00
X64OP_LOAD_AND_STORE, // lock and [mem],reg
};
2015-02-15 17:13:06 +00:00
void decode_x64_reg_op(const u8* code, x64_op_t& out_op, x64_reg_t& out_reg, size_t& out_size, size_t& out_length)
{
// simple analysis of x64 code allows to reinterpret MOV or other instructions in any desired way
2015-02-15 12:31:42 +00:00
out_length = 0;
u8 rex = 0, pg2 = 0;
2015-02-15 12:31:42 +00:00
bool oso = false, lock = false, repne = false, repe = false;
enum : u8
{
LOCK = 0xf0,
REPNE = 0xf2,
REPE = 0xf3,
};
// check prefixes:
2015-02-15 12:31:42 +00:00
for (;; code++, out_length++)
{
switch (const u8 prefix = *code)
{
2015-02-15 12:31:42 +00:00
case LOCK: // group 1
{
if (lock)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): LOCK prefix found twice", (size_t)code - out_length);
2015-02-15 12:31:42 +00:00
}
lock = true;
continue;
}
case REPNE: // group 1
{
if (repne)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): REPNE/REPNZ prefix found twice", (size_t)code - out_length);
2015-02-15 12:31:42 +00:00
}
repne = true;
continue;
}
case REPE: // group 1
{
if (repe)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): REP/REPE/REPZ prefix found twice", (size_t)code - out_length);
2015-02-15 12:31:42 +00:00
}
repe = true;
continue;
}
case 0x2e: // group 2
case 0x36:
case 0x3e:
case 0x26:
case 0x64:
case 0x65:
{
2015-02-15 12:31:42 +00:00
if (pg2)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): 0x%02x (group 2 prefix) found after 0x%02x", (size_t)code - out_length, prefix, pg2);
}
else
{
2015-02-15 12:31:42 +00:00
pg2 = prefix; // probably, segment register
}
2015-02-15 12:31:42 +00:00
continue;
}
2015-02-15 12:31:42 +00:00
case 0x66: // group 3
{
if (oso)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): operand-size override prefix found twice", (size_t)code - out_length);
2015-02-15 12:31:42 +00:00
}
oso = true;
continue;
}
case 0x67: // group 4
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): address-size override prefix found", (size_t)code - out_length, prefix);
2015-02-15 17:13:06 +00:00
out_op = X64OP_NONE;
2015-02-15 12:31:42 +00:00
out_reg = X64_NOT_SET;
out_size = 0;
out_length = 0;
return;
}
default:
{
if ((prefix & 0xf0) == 0x40) // check REX prefix
{
if (rex)
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "decode_x64_reg_op(%016llxh): 0x%02x (REX prefix) found after 0x%02x", (size_t)code - out_length, prefix, rex);
}
2015-02-15 12:31:42 +00:00
else
{
2015-02-15 12:31:42 +00:00
rex = prefix;
}
continue;
}
}
}
break;
}
2015-02-15 12:31:42 +00:00
auto get_modRM_reg = [](const u8* code, const u8 rex) -> x64_reg_t
{
2015-02-15 17:13:06 +00:00
return (x64_reg_t)(((*code & 0x38) >> 3 | (/* check REX.R bit */ rex & 4 ? 8 : 0)) + X64R_RAX);
2015-02-15 12:31:42 +00:00
};
auto get_modRM_reg_xmm = [](const u8* code, const u8 rex) -> x64_reg_t
{
2015-02-15 17:13:06 +00:00
return (x64_reg_t)(((*code & 0x38) >> 3 | (/* check REX.R bit */ rex & 4 ? 8 : 0)) + X64R_XMM0);
2015-02-15 12:31:42 +00:00
};
auto get_modRM_reg_lh = [](const u8* code) -> x64_reg_t
{
2015-02-15 17:13:06 +00:00
return (x64_reg_t)(((*code & 0x38) >> 3) + X64R_AL);
};
2015-02-15 17:13:06 +00:00
auto get_op_size = [](const u8 rex, const bool oso) -> size_t
2015-02-15 12:31:42 +00:00
{
return rex & 8 ? 8 : (oso ? 2 : 4);
};
2015-02-15 17:13:06 +00:00
auto get_modRM_size = [](const u8* code) -> size_t
{
switch (*code >> 6) // check Mod
{
case 0: return (*code & 0x07) == 4 ? 2 : 1; // check SIB
case 1: return (*code & 0x07) == 4 ? 3 : 2; // check SIB (disp8)
case 2: return (*code & 0x07) == 4 ? 6 : 5; // check SIB (disp32)
default: return 1;
}
};
2015-02-15 12:31:42 +00:00
const u8 op1 = (out_length++, *code++), op2 = code[0], op3 = code[1];
switch (op1)
{
case 0x0f:
{
out_length++, code++;
switch (op2)
{
case 0x7f:
{
2015-02-19 11:18:28 +00:00
if ((repe && !oso) || (!repe && oso)) // MOVDQU/MOVDQA xmm/m, xmm
2015-02-15 12:31:42 +00:00
{
out_op = X64OP_STORE;
out_reg = get_modRM_reg_xmm(code, rex);
out_size = 16;
out_length += get_modRM_size(code);
return;
}
break;
}
case 0xb0:
{
if (!oso) // CMPXCHG r8/m8, r8
{
out_op = X64OP_CMPXCHG;
out_reg = rex & 8 ? get_modRM_reg(code, rex) : get_modRM_reg_lh(code);
out_size = 1;
out_length += get_modRM_size(code);
return;
}
break;
}
case 0xb1:
{
if (true) // CMPXCHG r/m, r (16, 32, 64)
{
out_op = X64OP_CMPXCHG;
out_reg = get_modRM_reg(code, rex);
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code);
return;
}
break;
}
}
break;
}
2015-03-06 22:10:04 +00:00
case 0x20:
{
if (!oso)
{
out_op = X64OP_LOAD_AND_STORE;
out_reg = rex & 8 ? get_modRM_reg(code, rex) : get_modRM_reg_lh(code);
out_size = 1;
out_length += get_modRM_size(code);
return;
}
break;
}
case 0x21:
{
if (true)
{
out_op = X64OP_LOAD_AND_STORE;
out_reg = get_modRM_reg(code, rex);
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code);
return;
}
break;
}
2015-02-15 12:31:42 +00:00
case 0x86:
{
2015-02-15 12:31:42 +00:00
if (!oso) // XCHG r8/m8, r8
{
out_op = X64OP_XCHG;
out_reg = rex & 8 ? get_modRM_reg(code, rex) : get_modRM_reg_lh(code);
out_size = 1;
out_length += get_modRM_size(code);
return;
}
break;
}
case 0x87:
{
if (true) // XCHG r/m, r (16, 32, 64)
{
out_op = X64OP_XCHG;
out_reg = get_modRM_reg(code, rex);
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code);
return;
}
break;
}
case 0x88:
{
2015-02-15 12:31:42 +00:00
if (!lock && !oso) // MOV r8/m8, r8
{
out_op = X64OP_STORE;
out_reg = rex & 8 ? get_modRM_reg(code, rex) : get_modRM_reg_lh(code);
out_size = 1;
out_length += get_modRM_size(code);
return;
}
break;
}
2015-02-15 12:31:42 +00:00
case 0x89:
{
2015-02-15 12:31:42 +00:00
if (!lock) // MOV r/m, r (16, 32, 64)
{
out_op = X64OP_STORE;
out_reg = get_modRM_reg(code, rex);
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code);
return;
}
break;
}
2015-02-15 12:31:42 +00:00
case 0x8a:
{
if (!lock && !oso) // MOV r8, r8/m8
{
out_op = X64OP_LOAD;
out_reg = rex & 8 ? get_modRM_reg(code, rex) : get_modRM_reg_lh(code);
out_size = 1;
out_length += get_modRM_size(code);
return;
}
break;
}
case 0x8b:
{
if (!lock) // MOV r, r/m (16, 32, 64)
{
out_op = X64OP_LOAD;
out_reg = get_modRM_reg(code, rex);
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code);
return;
}
break;
}
case 0xa4:
{
if (!oso && !lock && !repe && !rex) // MOVS
{
out_op = X64OP_MOVS;
out_reg = X64_NOT_SET;
out_size = 1;
return;
}
if (!oso && !lock && repe) // REP MOVS
{
out_op = X64OP_MOVS;
out_reg = rex & 8 ? X64R_RCX : X64R_ECX;
out_size = 1;
return;
}
break;
}
2015-02-17 02:01:47 +00:00
case 0xaa:
{
if (!oso && !lock && !repe && !rex) // STOS
{
out_op = X64OP_STOS;
out_reg = X64_NOT_SET;
out_size = 1;
return;
}
if (!oso && !lock && repe) // REP STOS
{
out_op = X64OP_STOS;
out_reg = rex & 8 ? X64R_RCX : X64R_ECX;
out_size = 1;
return;
}
break;
}
2015-02-15 12:31:42 +00:00
case 0xc6:
{
2015-02-15 12:31:42 +00:00
if (!lock && !oso && get_modRM_reg(code, 0) == X64R_RAX) // MOV r8/m8, imm8
{
2015-02-15 12:31:42 +00:00
out_op = X64OP_STORE;
out_reg = X64_IMM8;
out_size = 1;
out_length += get_modRM_size(code) + 1;
return;
}
2015-02-15 12:31:42 +00:00
break;
}
2015-02-15 12:31:42 +00:00
case 0xc7:
{
2015-02-15 12:31:42 +00:00
if (!lock && get_modRM_reg(code, 0) == X64R_RAX) // MOV r/m, imm16/imm32 (16, 32, 64)
{
out_op = X64OP_STORE;
out_reg = oso ? X64_IMM16 : X64_IMM32;
out_size = get_op_size(rex, oso);
out_length += get_modRM_size(code) + (oso ? 2 : 4);
return;
}
break;
}
}
2015-02-15 12:31:42 +00:00
2015-02-17 00:08:23 +00:00
LOG_WARNING(MEMORY, "decode_x64_reg_op(%016llxh): unsupported opcode found (%016llX%016llX)", (size_t)code - out_length, *(be_t<u64>*)(code - out_length), *(be_t<u64>*)(code - out_length + 8));
2015-02-15 17:13:06 +00:00
out_op = X64OP_NONE;
2015-02-15 12:31:42 +00:00
out_reg = X64_NOT_SET;
out_size = 0;
out_length = 0;
}
#ifdef _WIN32
2015-01-18 13:57:39 +00:00
typedef CONTEXT x64_context;
2015-02-17 00:08:23 +00:00
#define X64REG(context, reg) (&(&(context)->Rax)[reg])
#define XMMREG(context, reg) (reinterpret_cast<u128*>(&(&(context)->Xmm0)[reg]))
#define EFLAGS(context) ((context)->EFlags)
2015-01-18 13:57:39 +00:00
#else
typedef ucontext_t x64_context;
2015-02-11 04:17:39 +00:00
#ifdef __APPLE__
#define X64REG(context, reg) (darwin_x64reg(context, reg))
2015-03-02 23:34:49 +00:00
#define XMMREG(context, reg) (reinterpret_cast<u128*>(&(context)->uc_mcontext->__fs.__fpu_xmm0.__xmm_reg[reg]))
2015-02-28 19:37:31 +00:00
#define EFLAGS(context) ((context)->uc_mcontext->__ss.__rflags)
2015-02-11 04:17:39 +00:00
uint64_t* darwin_x64reg(x64_context *context, int reg)
{
2015-02-13 20:24:18 +00:00
auto *state = &context->uc_mcontext->__ss;
switch(reg)
{
case 0: // RAX
return &state->__rax;
case 1: // RCX
return &state->__rcx;
case 2: // RDX
return &state->__rdx;
case 3: // RBX
return &state->__rbx;
case 4: // RSP
return &state->__rsp;
case 5: // RBP
return &state->__rbp;
case 6: // RSI
return &state->__rsi;
case 7: // RDI
return &state->__rdi;
case 8: // R8
return &state->__r8;
case 9: // R9
return &state->__r9;
case 10: // R10
return &state->__r10;
case 11: // R11
return &state->__r11;
case 12: // R12
return &state->__r12;
case 13: // R13
return &state->__r13;
case 14: // R14
return &state->__r14;
case 15: // R15
return &state->__r15;
case 16: // RIP
return &state->__rip;
default: // FAIL
assert(0);
}
2015-02-11 04:17:39 +00:00
}
#else
2015-02-13 20:24:18 +00:00
2015-02-11 04:17:39 +00:00
typedef decltype(REG_RIP) reg_table_t;
2015-01-18 13:57:39 +00:00
static const reg_table_t reg_table[17] =
{
2015-01-18 13:57:39 +00:00
REG_RAX, REG_RCX, REG_RDX, REG_RBX, REG_RSP, REG_RBP, REG_RSI, REG_RDI,
REG_R8, REG_R9, REG_R10, REG_R11, REG_R12, REG_R13, REG_R14, REG_R15, REG_RIP
};
2015-02-17 00:08:23 +00:00
#define X64REG(context, reg) (&(context)->uc_mcontext.gregs[reg_table[reg]])
#define XMMREG(context, reg) (reinterpret_cast<u128*>(&(context)->uc_mcontext.fpregs->_xmm[reg]))
2015-02-17 00:40:10 +00:00
#define EFLAGS(context) ((context)->uc_mcontext.gregs[REG_EFL])
2015-02-13 20:24:18 +00:00
2015-02-11 04:17:39 +00:00
#endif // __APPLE__
2015-01-18 13:57:39 +00:00
#endif
2015-02-15 17:13:06 +00:00
#define RAX(c) (*X64REG((c), 0))
#define RCX(c) (*X64REG((c), 1))
#define RDX(c) (*X64REG((c), 2))
#define RSI(c) (*X64REG((c), 6))
#define RDI(c) (*X64REG((c), 7))
#define RIP(c) (*X64REG((c), 16))
bool get_x64_reg_value(x64_context* context, x64_reg_t reg, size_t d_size, size_t i_size, u64& out_value)
{
// get x64 reg value (for store operations)
if (reg - X64R_RAX < 16)
{
// load the value from x64 register
const u64 reg_value = *X64REG(context, reg - X64R_RAX);
switch (d_size)
{
case 1: out_value = (u8)reg_value; return true;
case 2: out_value = (u16)reg_value; return true;
case 4: out_value = (u32)reg_value; return true;
case 8: out_value = reg_value; return true;
}
}
else if (reg - X64R_AL < 4 && d_size == 1)
{
out_value = (u8)(*X64REG(context, reg - X64R_AL));
2015-02-16 01:53:53 +00:00
return true;
2015-02-15 17:13:06 +00:00
}
else if (reg - X64R_AH < 4 && d_size == 1)
{
out_value = (u8)(*X64REG(context, reg - X64R_AH) >> 8);
2015-02-16 01:53:53 +00:00
return true;
2015-02-15 17:13:06 +00:00
}
2015-02-19 20:38:25 +00:00
else if (reg == X64_IMM8)
{
// load the immediate value (assuming it's at the end of the instruction)
2015-02-19 20:38:25 +00:00
const s8 imm_value = *(s8*)(RIP(context) + i_size - 1);
switch (d_size)
{
case 1: out_value = (u8)imm_value; return true;
}
}
else if (reg == X64_IMM16)
{
const s16 imm_value = *(s16*)(RIP(context) + i_size - 2);
switch (d_size)
{
case 2: out_value = (u16)imm_value; return true;
}
}
2015-02-15 17:13:06 +00:00
else if (reg == X64_IMM32)
{
const s32 imm_value = *(s32*)(RIP(context) + i_size - 4);
switch (d_size)
{
case 4: out_value = (u32)imm_value; return true;
case 8: out_value = (u64)imm_value; return true; // sign-extended
}
}
else if (reg == X64R_ECX)
{
out_value = (u32)RCX(context);
2015-02-16 01:53:53 +00:00
return true;
2015-02-15 17:13:06 +00:00
}
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "get_x64_reg_value(): invalid arguments (reg=%d, d_size=%lld, i_size=%lld)", reg, d_size, i_size);
2015-02-15 17:13:06 +00:00
return false;
}
bool put_x64_reg_value(x64_context* context, x64_reg_t reg, size_t d_size, u64 value)
{
// save x64 reg value (for load operations)
if (reg - X64R_RAX < 16)
{
2015-02-17 00:08:23 +00:00
// save the value into x64 register
switch (d_size)
{
case 1: *X64REG(context, reg - X64R_RAX) = value & 0xff | *X64REG(context, reg - X64R_RAX) & 0xffffff00; return true;
case 2: *X64REG(context, reg - X64R_RAX) = value & 0xffff | *X64REG(context, reg - X64R_RAX) & 0xffff0000; return true;
case 4: *X64REG(context, reg - X64R_RAX) = value & 0xffffffff; return true;
case 8: *X64REG(context, reg - X64R_RAX) = value; return true;
}
2015-02-15 17:13:06 +00:00
}
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "put_x64_reg_value(): invalid destination (reg=%d, d_size=%lld, value=0x%llx)", reg, d_size, value);
2015-02-15 17:13:06 +00:00
return false;
}
2015-02-17 00:08:23 +00:00
bool set_x64_cmp_flags(x64_context* context, size_t d_size, u64 x, u64 y)
{
switch (d_size)
{
case 1: break;
case 2: break;
case 4: break;
case 8: break;
default: LOG_ERROR(MEMORY, "set_x64_cmp_flags(): invalid d_size (%lld)", d_size); return false;
}
const u64 sign = 1ull << (d_size * 8 - 1); // sign mask
const u64 diff = x - y;
const u64 summ = x + y;
if (((x & y) | ((x ^ y) & ~summ)) & sign)
{
EFLAGS(context) |= 0x1; // set CF
}
else
{
EFLAGS(context) &= ~0x1; // clear CF
}
if (x == y)
{
EFLAGS(context) |= 0x40; // set ZF
}
else
{
EFLAGS(context) &= ~0x40; // clear ZF
}
if (diff & sign)
{
EFLAGS(context) |= 0x80; // set SF
}
else
{
EFLAGS(context) &= ~0x80; // clear SF
}
if ((x ^ summ) & (y ^ summ) & sign)
{
EFLAGS(context) |= 0x800; // set OF
}
else
{
EFLAGS(context) &= ~0x800; // clear OF
}
const u8 p1 = (u8)diff ^ ((u8)diff >> 4);
const u8 p2 = p1 ^ (p1 >> 2);
const u8 p3 = p2 ^ (p2 >> 1);
if ((p3 & 1) == 0)
{
EFLAGS(context) |= 0x4; // set PF
}
else
{
EFLAGS(context) &= ~0x4; // clear PF
}
if (((x & y) | ((x ^ y) & ~summ)) & 0x8)
{
EFLAGS(context) |= 0x10; // set AF
}
else
{
EFLAGS(context) &= ~0x10; // clear AF
}
return true;
}
size_t get_x64_access_size(x64_context* context, x64_op_t op, x64_reg_t reg, size_t d_size, size_t i_size)
2015-02-15 17:13:06 +00:00
{
2015-02-17 15:27:15 +00:00
if (op == X64OP_MOVS || op == X64OP_STOS)
2015-02-15 17:13:06 +00:00
{
2015-02-17 15:27:15 +00:00
if (EFLAGS(context) & 0x400 /* direction flag */)
2015-02-15 17:13:06 +00:00
{
2015-02-17 15:27:15 +00:00
// skip reservation bound check (TODO)
return 0;
2015-02-15 17:13:06 +00:00
}
2015-02-17 15:27:15 +00:00
if (reg != X64_NOT_SET) // get "full" access size from RCX register
{
u64 counter;
if (!get_x64_reg_value(context, reg, 8, i_size, counter))
{
return -1;
}
return d_size * counter;
}
}
if (op == X64OP_CMPXCHG)
{
// detect whether this instruction can't actually modify memory to avoid breaking reservation;
// this may theoretically cause endless loop, but it shouldn't be a problem if only load_sync() generates such instruction
2015-02-17 15:27:15 +00:00
u64 cmp, exch;
if (!get_x64_reg_value(context, reg, d_size, i_size, cmp) || !get_x64_reg_value(context, X64R_RAX, d_size, i_size, exch))
{
return -1;
}
if (cmp == exch)
{
// skip reservation bound check
return 0;
}
2015-02-15 17:13:06 +00:00
}
2015-02-17 00:08:23 +00:00
return d_size;
2015-02-15 17:13:06 +00:00
}
/**
* Callback that can be customised by GSRender backends to track memory access.
* Backends can protect memory pages and get this callback called when an access
* violation is met.
* Should return true if the backend handles the access violation.
*/
std::function<bool(u32 addr)> gfxHandler = [](u32) { return false; };
2015-02-17 00:08:23 +00:00
bool handle_access_violation(u32 addr, bool is_writing, x64_context* context)
2015-01-18 13:57:39 +00:00
{
2015-02-15 17:13:06 +00:00
auto code = (const u8*)RIP(context);
2015-02-15 12:31:42 +00:00
x64_op_t op;
x64_reg_t reg;
2015-02-15 17:13:06 +00:00
size_t d_size;
size_t i_size;
2015-02-15 12:31:42 +00:00
if (gfxHandler(addr))
return true;
2015-02-15 12:31:42 +00:00
// decode single x64 instruction that causes memory access
2015-02-15 17:13:06 +00:00
decode_x64_reg_op(code, op, reg, d_size, i_size);
2015-02-15 12:31:42 +00:00
2015-02-17 00:08:23 +00:00
if ((d_size | d_size + addr) >= 0x100000000ull)
2015-02-16 01:53:53 +00:00
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "Invalid d_size (0x%llx)", d_size);
2015-02-16 01:53:53 +00:00
return false;
}
2015-02-17 00:08:23 +00:00
// get length of data being accessed
size_t a_size = get_x64_access_size(context, op, reg, d_size, i_size);
2015-02-16 01:53:53 +00:00
2015-02-17 00:08:23 +00:00
if ((a_size | a_size + addr) >= 0x100000000ull)
{
LOG_ERROR(MEMORY, "Invalid a_size (0x%llx)", a_size);
return false;
2015-02-16 01:53:53 +00:00
}
2015-02-06 23:39:51 +00:00
// check if address is RawSPU MMIO register
if (addr - RAW_SPU_BASE_ADDR < (6 * RAW_SPU_OFFSET) && (addr % RAW_SPU_OFFSET) >= RAW_SPU_PROB_OFFSET)
{
2015-03-04 21:51:14 +00:00
auto t = Emu.GetCPU().GetRawSPUThread((addr - RAW_SPU_BASE_ADDR) / RAW_SPU_OFFSET);
if (!t)
{
return false;
}
2015-02-17 00:08:23 +00:00
if (a_size != 4 || !d_size || !i_size)
2015-02-15 12:31:42 +00:00
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "Invalid or unsupported instruction (op=%d, reg=%d, d_size=%lld, a_size=0x%llx, i_size=%lld)", op, reg, d_size, a_size, i_size);
2015-02-15 12:31:42 +00:00
return false;
2015-01-18 13:57:39 +00:00
}
2015-03-04 21:51:14 +00:00
auto& spu = static_cast<RawSPUThread&>(*t);
2015-01-18 13:57:39 +00:00
switch (op)
{
case X64OP_LOAD:
{
2015-02-15 17:13:06 +00:00
u32 value;
if (is_writing || !spu.ReadReg(addr, value) || !put_x64_reg_value(context, reg, d_size, _byteswap_ulong(value)))
2015-02-15 17:13:06 +00:00
{
return false;
}
2015-01-18 13:57:39 +00:00
break;
}
case X64OP_STORE:
{
2015-02-15 17:13:06 +00:00
u64 reg_value;
if (!is_writing || !get_x64_reg_value(context, reg, d_size, i_size, reg_value) || !spu.WriteReg(addr, _byteswap_ulong((u32)reg_value)))
2015-02-15 17:13:06 +00:00
{
return false;
}
2015-01-18 13:57:39 +00:00
break;
}
2015-02-17 02:01:47 +00:00
case X64OP_MOVS: // possibly, TODO
case X64OP_STOS:
2015-02-15 12:31:42 +00:00
default:
{
2015-02-17 00:08:23 +00:00
LOG_ERROR(MEMORY, "Invalid or unsupported operation (op=%d, reg=%d, d_size=%lld, i_size=%lld)", op, reg, d_size, i_size);
2015-02-15 12:31:42 +00:00
return false;
}
2015-01-18 13:57:39 +00:00
}
2015-02-15 17:13:06 +00:00
// skip processed instruction
RIP(context) += i_size;
2015-01-18 13:57:39 +00:00
return true;
}
2015-02-17 00:08:23 +00:00
// check if fault is caused by the reservation
return vm::reservation_query(addr, (u32)a_size, is_writing, [&]() -> bool
{
// write memory using "privileged" access to avoid breaking reservation
if (!d_size || !i_size)
{
LOG_ERROR(MEMORY, "Invalid or unsupported instruction (op=%d, reg=%d, d_size=%lld, a_size=0x%llx, i_size=%lld)", op, reg, d_size, a_size, i_size);
return false;
}
switch (op)
{
case X64OP_STORE:
{
if (d_size == 16)
{
if (reg - X64R_XMM0 >= 16)
{
LOG_ERROR(MEMORY, "X64OP_STORE: d_size=16, reg=%d", reg);
return false;
}
2015-02-28 14:41:15 +00:00
memcpy(vm::priv_ptr(addr), XMMREG(context, reg - X64R_XMM0), 16);
2015-02-17 00:08:23 +00:00
break;
}
u64 reg_value;
if (!get_x64_reg_value(context, reg, d_size, i_size, reg_value))
{
return false;
}
2015-02-28 14:41:15 +00:00
memcpy(vm::priv_ptr(addr), &reg_value, d_size);
2015-02-17 00:08:23 +00:00
break;
}
case X64OP_MOVS:
{
if (d_size > 8)
{
LOG_ERROR(MEMORY, "X64OP_MOVS: d_size=%lld", d_size);
return false;
}
if (vm::get_ptr(addr) != (void*)RDI(context))
{
2015-02-17 15:27:15 +00:00
LOG_ERROR(MEMORY, "X64OP_MOVS: rdi=0x%llx, rsi=0x%llx, addr=0x%x", (u64)RDI(context), (u64)RSI(context), addr);
2015-02-17 00:08:23 +00:00
return false;
}
u32 a_addr = addr;
while (a_addr >> 12 == addr >> 12)
{
u64 value;
// copy data
memcpy(&value, (void*)RSI(context), d_size);
2015-02-28 14:41:15 +00:00
memcpy(vm::priv_ptr(a_addr), &value, d_size);
2015-02-17 00:08:23 +00:00
// shift pointers
if (EFLAGS(context) & 0x400 /* direction flag */)
{
LOG_ERROR(MEMORY, "X64OP_MOVS TODO: reversed direction");
return false;
//RSI(context) -= d_size;
//RDI(context) -= d_size;
//a_addr -= (u32)d_size;
}
else
{
RSI(context) += d_size;
RDI(context) += d_size;
a_addr += (u32)d_size;
}
// decrement counter
if (reg == X64_NOT_SET || !--RCX(context))
{
break;
}
}
if (reg == X64_NOT_SET || !RCX(context))
{
break;
}
// don't skip partially processed instruction
return true;
}
2015-02-17 02:01:47 +00:00
case X64OP_STOS:
{
if (d_size > 8)
{
LOG_ERROR(MEMORY, "X64OP_STOS: d_size=%lld", d_size);
return false;
}
if (vm::get_ptr(addr) != (void*)RDI(context))
{
2015-02-17 15:27:15 +00:00
LOG_ERROR(MEMORY, "X64OP_STOS: rdi=0x%llx, addr=0x%x", (u64)RDI(context), addr);
2015-02-17 02:01:47 +00:00
return false;
}
u64 value;
if (!get_x64_reg_value(context, X64R_RAX, d_size, i_size, value))
{
return false;
}
u32 a_addr = addr;
while (a_addr >> 12 == addr >> 12)
{
// fill data with value
2015-02-28 14:41:15 +00:00
memcpy(vm::priv_ptr(a_addr), &value, d_size);
2015-02-17 02:01:47 +00:00
// shift pointers
if (EFLAGS(context) & 0x400 /* direction flag */)
{
LOG_ERROR(MEMORY, "X64OP_STOS TODO: reversed direction");
return false;
//RDI(context) -= d_size;
//a_addr -= (u32)d_size;
}
else
{
RDI(context) += d_size;
a_addr += (u32)d_size;
}
// decrement counter
if (reg == X64_NOT_SET || !--RCX(context))
{
break;
}
}
if (reg == X64_NOT_SET || !RCX(context))
{
break;
}
// don't skip partially processed instruction
return true;
}
2015-02-17 00:08:23 +00:00
case X64OP_XCHG:
{
u64 reg_value;
if (!get_x64_reg_value(context, reg, d_size, i_size, reg_value))
{
return false;
}
switch (d_size)
{
case 1: reg_value = vm::priv_ref<atomic<u8>>(addr).exchange((u8)reg_value); break;
case 2: reg_value = vm::priv_ref<atomic<u16>>(addr).exchange((u16)reg_value); break;
case 4: reg_value = vm::priv_ref<atomic<u32>>(addr).exchange((u32)reg_value); break;
case 8: reg_value = vm::priv_ref<atomic<u64>>(addr).exchange((u64)reg_value); break;
2015-02-17 15:27:15 +00:00
default: return false;
2015-02-17 00:08:23 +00:00
}
if (!put_x64_reg_value(context, reg, d_size, reg_value))
{
return false;
}
break;
}
case X64OP_CMPXCHG:
{
u64 reg_value, old_value, cmp_value;
if (!get_x64_reg_value(context, reg, d_size, i_size, reg_value) || !get_x64_reg_value(context, X64R_RAX, d_size, i_size, cmp_value))
{
return false;
}
switch (d_size)
{
case 1: old_value = vm::priv_ref<atomic<u8>>(addr).compare_and_swap((u8)cmp_value, (u8)reg_value); break;
case 2: old_value = vm::priv_ref<atomic<u16>>(addr).compare_and_swap((u16)cmp_value, (u16)reg_value); break;
case 4: old_value = vm::priv_ref<atomic<u32>>(addr).compare_and_swap((u32)cmp_value, (u32)reg_value); break;
case 8: old_value = vm::priv_ref<atomic<u64>>(addr).compare_and_swap((u64)cmp_value, (u64)reg_value); break;
2015-02-17 15:27:15 +00:00
default: return false;
2015-02-17 00:08:23 +00:00
}
if (!put_x64_reg_value(context, X64R_RAX, d_size, old_value) || !set_x64_cmp_flags(context, d_size, cmp_value, old_value))
{
return false;
}
break;
}
2015-03-06 22:10:04 +00:00
case X64OP_LOAD_AND_STORE:
{
u64 value;
if (!get_x64_reg_value(context, reg, d_size, i_size, value))
{
return false;
}
switch (d_size)
{
case 1: value = vm::priv_ref<atomic<u8>>(addr) &= static_cast<u8>(value); break;
case 2: value = vm::priv_ref<atomic<u16>>(addr) &= static_cast<u16>(value); break;
case 4: value = vm::priv_ref<atomic<u32>>(addr) &= static_cast<u32>(value); break;
case 8: value = vm::priv_ref<atomic<u64>>(addr) &= value; break;
2015-03-06 22:10:04 +00:00
default: return false;
}
if (!set_x64_cmp_flags(context, d_size, value, 0))
{
return false;
}
break;
}
2015-02-17 00:08:23 +00:00
default:
{
LOG_ERROR(MEMORY, "Invalid or unsupported operation (op=%d, reg=%d, d_size=%lld, a_size=0x%llx, i_size=%lld)", op, reg, d_size, a_size, i_size);
return false;
}
}
// skip processed instruction
RIP(context) += i_size;
return true;
});
2015-01-18 13:57:39 +00:00
// TODO: allow recovering from a page fault as a feature of PS3 virtual memory
}
2015-01-18 13:57:39 +00:00
#ifdef _WIN32
2015-01-18 13:57:39 +00:00
void _se_translator(unsigned int u, EXCEPTION_POINTERS* pExp)
{
2015-02-08 13:38:08 +00:00
const u64 addr64 = (u64)pExp->ExceptionRecord->ExceptionInformation[1] - (u64)vm::g_base_addr;
2015-01-18 13:57:39 +00:00
const bool is_writing = pExp->ExceptionRecord->ExceptionInformation[0] != 0;
2015-02-08 13:38:08 +00:00
2015-01-18 13:57:39 +00:00
if (u == EXCEPTION_ACCESS_VIOLATION && (u32)addr64 == addr64)
{
throw fmt::format("Access violation %s location 0x%llx", is_writing ? "writing" : "reading", addr64);
}
}
2015-02-08 13:38:08 +00:00
const PVOID exception_handler = (atexit([]{ RemoveVectoredExceptionHandler(exception_handler); }), AddVectoredExceptionHandler(1, [](PEXCEPTION_POINTERS pExp) -> LONG
2015-02-06 23:39:51 +00:00
{
2015-02-08 13:38:08 +00:00
const u64 addr64 = (u64)pExp->ExceptionRecord->ExceptionInformation[1] - (u64)vm::g_base_addr;
2015-02-08 15:25:50 +00:00
const bool is_writing = pExp->ExceptionRecord->ExceptionInformation[0] != 0;
2015-02-06 23:39:51 +00:00
2015-02-08 13:38:08 +00:00
if (pExp->ExceptionRecord->ExceptionCode == EXCEPTION_ACCESS_VIOLATION &&
(u32)addr64 == addr64 &&
2015-06-30 22:25:52 +00:00
get_current_thread_ctrl() &&
2015-02-08 15:25:50 +00:00
handle_access_violation((u32)addr64, is_writing, pExp->ContextRecord))
2015-02-06 23:39:51 +00:00
{
2015-02-08 13:38:08 +00:00
return EXCEPTION_CONTINUE_EXECUTION;
2015-02-06 23:39:51 +00:00
}
else
{
return EXCEPTION_CONTINUE_SEARCH;
}
2015-02-08 13:38:08 +00:00
}));
2015-02-06 23:39:51 +00:00
2015-06-30 22:25:52 +00:00
const auto exception_filter = SetUnhandledExceptionFilter([](PEXCEPTION_POINTERS pExp) -> LONG
{
_se_translator(pExp->ExceptionRecord->ExceptionCode, pExp);
return EXCEPTION_CONTINUE_SEARCH;
});
#else
2015-01-04 22:29:51 +00:00
void signal_handler(int sig, siginfo_t* info, void* uct)
{
2015-02-08 13:38:08 +00:00
const u64 addr64 = (u64)info->si_addr - (u64)vm::g_base_addr;
2015-02-11 04:17:39 +00:00
#ifdef __APPLE__
2015-02-13 20:24:18 +00:00
const bool is_writing = ((ucontext_t*)uct)->uc_mcontext->__es.__err & 0x2;
2015-02-11 04:17:39 +00:00
#else
2015-02-08 13:38:08 +00:00
const bool is_writing = ((ucontext_t*)uct)->uc_mcontext.gregs[REG_ERR] & 0x2;
2015-02-11 04:17:39 +00:00
#endif
2015-02-08 13:38:08 +00:00
2015-06-30 22:25:52 +00:00
if ((u32)addr64 == addr64 && get_current_thread_ctrl())
{
2015-02-08 15:25:50 +00:00
if (handle_access_violation((u32)addr64, is_writing, (ucontext_t*)uct))
{
2015-01-18 13:57:39 +00:00
return; // proceed execution
}
2015-01-18 13:57:39 +00:00
// TODO: this may be wrong
2015-02-08 13:38:08 +00:00
throw fmt::format("Access violation %s location 0x%llx", is_writing ? "writing" : "reading", addr64);
}
2015-01-08 22:17:26 +00:00
// else some fatal error
exit(EXIT_FAILURE);
}
const int sigaction_result = []() -> int
{
struct sigaction sa;
sa.sa_flags = SA_SIGINFO;
sigemptyset(&sa.sa_mask);
sa.sa_sigaction = signal_handler;
return sigaction(SIGSEGV, &sa, NULL);
2015-01-04 22:29:51 +00:00
}();
#endif
2015-06-30 22:25:52 +00:00
thread_local thread_ctrl_t* g_tls_this_thread = nullptr;
2015-06-30 22:25:52 +00:00
const thread_ctrl_t* get_current_thread_ctrl()
{
return g_tls_this_thread;
}
2015-06-30 22:25:52 +00:00
std::string thread_ctrl_t::get_name() const
{
return name();
}
void thread_ctrl_t::set_current()
2014-08-20 14:23:48 +00:00
{
2015-01-16 14:36:53 +00:00
const auto old_value = g_tls_this_thread;
2014-09-14 22:17:24 +00:00
2015-06-30 22:25:52 +00:00
if (old_value == this)
2014-09-14 22:17:24 +00:00
{
return;
}
2015-02-06 23:39:51 +00:00
if (old_value)
{
vm::reservation_free();
}
2015-06-30 22:25:52 +00:00
if (true && assigned.exchange(true))
2014-09-14 22:17:24 +00:00
{
2015-06-30 22:25:52 +00:00
LOG_ERROR(GENERAL, "Thread '%s' was already assigned to g_tls_this_thread of another thread", get_name());
2014-09-14 22:17:24 +00:00
g_tls_this_thread = nullptr;
}
else
{
2015-06-30 22:25:52 +00:00
g_tls_this_thread = this;
2014-09-14 22:17:24 +00:00
}
if (old_value)
{
2015-06-30 22:25:52 +00:00
old_value->assigned = false;
2014-09-14 22:17:24 +00:00
}
2014-08-20 14:23:48 +00:00
}
2015-06-30 22:25:52 +00:00
thread_t::thread_t(std::function<std::string()> name, std::function<void()> func)
2014-08-25 18:09:48 +00:00
{
2015-06-30 22:25:52 +00:00
start(std::move(name), func);
2014-08-25 18:09:48 +00:00
}
2015-06-30 22:25:52 +00:00
thread_t::~thread_t()
{
2015-06-30 22:25:52 +00:00
if (m_thread)
2014-08-20 14:23:48 +00:00
{
2015-06-30 22:25:52 +00:00
if (g_tls_this_thread != m_thread.get())
2015-02-08 13:38:08 +00:00
{
2015-06-30 22:25:52 +00:00
m_thread->m_thread.join();
2015-02-08 13:38:08 +00:00
}
2015-06-30 22:25:52 +00:00
else
2014-09-11 19:18:19 +00:00
{
2015-06-30 22:25:52 +00:00
m_thread->m_thread.detach();
2014-09-11 19:18:19 +00:00
}
2015-06-30 22:25:52 +00:00
}
}
2015-06-30 22:25:52 +00:00
std::string thread_t::get_name() const
{
2015-06-30 22:25:52 +00:00
if (!m_thread)
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Invalid thread");
}
2015-06-30 22:25:52 +00:00
if (!m_thread->name)
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Invalid name getter");
}
2015-06-30 22:25:52 +00:00
return m_thread->name();
}
2015-06-30 22:25:52 +00:00
std::atomic<u32> g_thread_count{ 0 };
2015-06-30 22:25:52 +00:00
void thread_t::start(std::function<std::string()> name, std::function<void()> func)
{
2015-06-30 22:25:52 +00:00
if (m_thread)
{
throw EXCEPTION("Thread already exists");
}
2015-06-30 22:25:52 +00:00
// create new ctrl and assign it
auto ctrl = std::make_shared<thread_ctrl_t>(std::move(name));
2015-06-30 22:25:52 +00:00
// start thread
ctrl->m_thread = std::thread([ctrl, func]()
{
g_thread_count++;
2015-06-30 22:25:52 +00:00
SetCurrentThreadDebugName(ctrl->get_name().c_str());
2015-06-30 22:25:52 +00:00
#if defined(_MSC_VER)
auto old_se_translator = _set_se_translator(_se_translator);
#endif
2014-02-27 18:25:32 +00:00
2015-06-30 22:25:52 +00:00
#ifdef _WIN32
if (!exception_handler || !exception_filter)
{
2015-06-30 22:25:52 +00:00
LOG_ERROR(GENERAL, "exception_handler not set");
return;
}
2015-06-30 22:25:52 +00:00
#else
if (sigaction_result == -1)
{
2015-06-30 22:25:52 +00:00
printf("sigaction() failed");
exit(EXIT_FAILURE);
}
#endif
2015-06-30 22:25:52 +00:00
// error handler
const auto error = [&](const char* text)
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
log_message(GENERAL, Emu.IsStopped() ? Log::Severity::Warning : Log::Severity::Error, "Exception: %s", text);
Emu.Pause();
};
2015-01-16 14:36:53 +00:00
2014-09-11 19:18:19 +00:00
try
{
2015-06-30 22:25:52 +00:00
ctrl->set_current();
if (Ini.HLELogging.GetValue())
{
LOG_NOTICE(GENERAL, "Thread started");
}
2014-09-11 19:18:19 +00:00
func();
}
catch (const char* e)
{
2015-06-30 22:25:52 +00:00
error(e);
2014-09-11 19:18:19 +00:00
}
catch (const std::string& e)
{
2015-06-30 22:25:52 +00:00
error(e.c_str());
}
catch (const fmt::exception& e)
{
error(e);
2014-09-11 19:18:19 +00:00
}
2015-01-16 14:36:53 +00:00
if (Emu.IsStopped())
{
2015-06-30 22:25:52 +00:00
LOG_NOTICE(GENERAL, "Thread aborted");
2015-01-16 14:36:53 +00:00
}
else if (Ini.HLELogging.GetValue())
{
2015-06-30 22:25:52 +00:00
LOG_NOTICE(GENERAL, "Thread ended");
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
//ctrl->set_current(false);
g_thread_count--;
2015-06-30 22:25:52 +00:00
ctrl->joinable = false;
ctrl->join_cv.notify_all();
#if defined(_MSC_VER)
_set_se_translator(old_se_translator);
#endif
2014-02-19 17:27:52 +00:00
});
2015-01-16 14:36:53 +00:00
2015-06-30 22:25:52 +00:00
// set
m_thread = std::move(ctrl);
}
void thread_t::detach()
{
if (!m_thread)
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Invalid thread");
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
const auto ctrl = std::move(m_thread);
ctrl->m_thread.detach();
}
2015-06-30 22:25:52 +00:00
void thread_t::join(std::unique_lock<std::mutex>& lock)
{
2015-06-30 22:25:52 +00:00
if (!m_thread)
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Invalid thread");
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
if (g_tls_this_thread == m_thread.get())
{
throw EXCEPTION("Deadlock");
}
const auto ctrl = std::move(m_thread);
// wait for completion
while (ctrl->joinable)
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
CHECK_EMU_STATUS;
ctrl->join_cv.wait_for(lock, std::chrono::milliseconds(1));
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
ctrl->m_thread.join();
}
2015-01-16 14:36:53 +00:00
void thread_t::join()
{
2015-06-30 22:25:52 +00:00
if (!m_thread)
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Invalid thread");
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
if (g_tls_this_thread == m_thread.get())
2015-01-16 14:36:53 +00:00
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Deadlock");
2015-01-16 14:36:53 +00:00
}
2015-06-30 22:25:52 +00:00
const auto ctrl = std::move(m_thread);
ctrl->m_thread.join();
}
2015-06-30 22:25:52 +00:00
bool thread_t::is_current() const
{
2015-06-30 22:25:52 +00:00
if (!m_thread)
{
throw EXCEPTION("Invalid thread");
}
return g_tls_this_thread == m_thread.get();
}
2015-06-30 22:25:52 +00:00
void waiter_map_t::check_emu_status(u32 addr)
{
if (Emu.IsStopped())
{
2015-06-30 22:25:52 +00:00
throw EXCEPTION("Aborted (emulation stopped) (%s, addr=0x%x)", name, addr);
}
}
void waiter_map_t::notify(u32 addr)
{
2015-06-30 22:25:52 +00:00
// signal an appropriate condition variable
cvs[get_hash(addr)].notify_all();
}
2015-01-17 16:36:23 +00:00
const std::function<bool()> SQUEUE_ALWAYS_EXIT = [](){ return true; };
const std::function<bool()> SQUEUE_NEVER_EXIT = [](){ return false; };
2015-01-16 17:09:53 +00:00
bool squeue_test_exit()
2014-12-25 20:30:34 +00:00
{
2015-01-16 17:09:53 +00:00
return Emu.IsStopped();
2014-12-25 20:30:34 +00:00
}