CPU/CodeCache: Fix crash on Apple Silicon
This commit is contained in:
parent
06c4dc5e1b
commit
f786138175
|
@ -18,6 +18,11 @@
|
||||||
#include <unistd.h>
|
#include <unistd.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#if defined(__APPLE__) && defined(__aarch64__)
|
||||||
|
// pthread_jit_write_protect_np()
|
||||||
|
#include <pthread.h>
|
||||||
|
#endif
|
||||||
|
|
||||||
Log_SetChannel(MemoryArena);
|
Log_SetChannel(MemoryArena);
|
||||||
|
|
||||||
#ifdef _WIN32
|
#ifdef _WIN32
|
||||||
|
@ -398,3 +403,31 @@ bool SharedMemoryMappingArea::Unmap(void* map_base, size_t map_size)
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#if defined(__APPLE__) && defined(__aarch64__)
|
||||||
|
|
||||||
|
static thread_local int s_code_write_depth = 0;
|
||||||
|
|
||||||
|
void MemMap::BeginCodeWrite()
|
||||||
|
{
|
||||||
|
// Log_DebugFmt("BeginCodeWrite(): {}", s_code_write_depth);
|
||||||
|
if ((s_code_write_depth++) == 0)
|
||||||
|
{
|
||||||
|
// Log_DebugPrint(" pthread_jit_write_protect_np(0)");
|
||||||
|
pthread_jit_write_protect_np(0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void MemMap::EndCodeWrite()
|
||||||
|
{
|
||||||
|
// Log_DebugFmt("EndCodeWrite(): {}", s_code_write_depth);
|
||||||
|
|
||||||
|
DebugAssert(s_code_write_depth > 0);
|
||||||
|
if ((--s_code_write_depth) == 0)
|
||||||
|
{
|
||||||
|
// Log_DebugPrint(" pthread_jit_write_protect_np(1)");
|
||||||
|
pthread_jit_write_protect_np(1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#endif
|
||||||
|
|
|
@ -41,6 +41,17 @@ void DestroySharedMemory(void* ptr);
|
||||||
void* MapSharedMemory(void* handle, size_t offset, void* baseaddr, size_t size, PageProtect mode);
|
void* MapSharedMemory(void* handle, size_t offset, void* baseaddr, size_t size, PageProtect mode);
|
||||||
void UnmapSharedMemory(void* baseaddr, size_t size);
|
void UnmapSharedMemory(void* baseaddr, size_t size);
|
||||||
bool MemProtect(void* baseaddr, size_t size, PageProtect mode);
|
bool MemProtect(void* baseaddr, size_t size, PageProtect mode);
|
||||||
|
|
||||||
|
/// JIT write protect for Apple Silicon. Needs to be called prior to writing to any RWX pages.
|
||||||
|
#if !defined(__APPLE__) || !defined(__aarch64__)
|
||||||
|
// clang-format off
|
||||||
|
ALWAYS_INLINE static void BeginCodeWrite() { }
|
||||||
|
ALWAYS_INLINE static void EndCodeWrite() { }
|
||||||
|
// clang-format on
|
||||||
|
#else
|
||||||
|
void BeginCodeWrite();
|
||||||
|
void EndCodeWrite();
|
||||||
|
#endif
|
||||||
} // namespace MemMap
|
} // namespace MemMap
|
||||||
|
|
||||||
class SharedMemoryMappingArea
|
class SharedMemoryMappingArea
|
||||||
|
|
|
@ -14,6 +14,7 @@
|
||||||
#include "common/assert.h"
|
#include "common/assert.h"
|
||||||
#include "common/intrin.h"
|
#include "common/intrin.h"
|
||||||
#include "common/log.h"
|
#include "common/log.h"
|
||||||
|
#include "common/memmap.h"
|
||||||
|
|
||||||
Log_SetChannel(CPU::CodeCache);
|
Log_SetChannel(CPU::CodeCache);
|
||||||
|
|
||||||
|
@ -628,6 +629,11 @@ void CPU::CodeCache::InvalidateBlocksWithPageIndex(u32 index)
|
||||||
new_block_state = BlockState::NeedsRecompile;
|
new_block_state = BlockState::NeedsRecompile;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!ppi.first_block_in_page)
|
||||||
|
return;
|
||||||
|
|
||||||
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
Block* block = ppi.first_block_in_page;
|
Block* block = ppi.first_block_in_page;
|
||||||
while (block)
|
while (block)
|
||||||
{
|
{
|
||||||
|
@ -637,6 +643,8 @@ void CPU::CodeCache::InvalidateBlocksWithPageIndex(u32 index)
|
||||||
|
|
||||||
ppi.first_block_in_page = nullptr;
|
ppi.first_block_in_page = nullptr;
|
||||||
ppi.last_block_in_page = nullptr;
|
ppi.last_block_in_page = nullptr;
|
||||||
|
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
CPU::CodeCache::PageProtectionMode CPU::CodeCache::GetProtectionModeForPC(u32 pc)
|
CPU::CodeCache::PageProtectionMode CPU::CodeCache::GetProtectionModeForPC(u32 pc)
|
||||||
|
@ -1279,6 +1287,7 @@ void CPU::CodeCache::CompileOrRevalidateBlock(u32 start_pc)
|
||||||
{
|
{
|
||||||
// TODO: this doesn't currently handle when the cache overflows...
|
// TODO: this doesn't currently handle when the cache overflows...
|
||||||
DebugAssert(IsUsingAnyRecompiler());
|
DebugAssert(IsUsingAnyRecompiler());
|
||||||
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
Block* block = LookupBlock(start_pc);
|
Block* block = LookupBlock(start_pc);
|
||||||
if (block)
|
if (block)
|
||||||
|
@ -1290,6 +1299,7 @@ void CPU::CodeCache::CompileOrRevalidateBlock(u32 start_pc)
|
||||||
DebugAssert(block->host_code);
|
DebugAssert(block->host_code);
|
||||||
SetCodeLUT(start_pc, block->host_code);
|
SetCodeLUT(start_pc, block->host_code);
|
||||||
BacklinkBlocks(start_pc, block->host_code);
|
BacklinkBlocks(start_pc, block->host_code);
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1303,6 +1313,7 @@ void CPU::CodeCache::CompileOrRevalidateBlock(u32 start_pc)
|
||||||
Log_ErrorFmt("Failed to read block at 0x{:08X}, falling back to uncached interpreter", start_pc);
|
Log_ErrorFmt("Failed to read block at 0x{:08X}, falling back to uncached interpreter", start_pc);
|
||||||
SetCodeLUT(start_pc, g_interpret_block);
|
SetCodeLUT(start_pc, g_interpret_block);
|
||||||
BacklinkBlocks(start_pc, g_interpret_block);
|
BacklinkBlocks(start_pc, g_interpret_block);
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1322,20 +1333,26 @@ void CPU::CodeCache::CompileOrRevalidateBlock(u32 start_pc)
|
||||||
Log_ErrorFmt("Failed to compile block at 0x{:08X}, falling back to uncached interpreter", start_pc);
|
Log_ErrorFmt("Failed to compile block at 0x{:08X}, falling back to uncached interpreter", start_pc);
|
||||||
SetCodeLUT(start_pc, g_interpret_block);
|
SetCodeLUT(start_pc, g_interpret_block);
|
||||||
BacklinkBlocks(start_pc, g_interpret_block);
|
BacklinkBlocks(start_pc, g_interpret_block);
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
SetCodeLUT(start_pc, block->host_code);
|
SetCodeLUT(start_pc, block->host_code);
|
||||||
BacklinkBlocks(start_pc, block->host_code);
|
BacklinkBlocks(start_pc, block->host_code);
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
void CPU::CodeCache::DiscardAndRecompileBlock(u32 start_pc)
|
void CPU::CodeCache::DiscardAndRecompileBlock(u32 start_pc)
|
||||||
{
|
{
|
||||||
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
Log_DevPrintf("Discard block %08X with manual protection", start_pc);
|
Log_DevPrintf("Discard block %08X with manual protection", start_pc);
|
||||||
Block* block = LookupBlock(start_pc);
|
Block* block = LookupBlock(start_pc);
|
||||||
DebugAssert(block && block->state == BlockState::Valid);
|
DebugAssert(block && block->state == BlockState::Valid);
|
||||||
InvalidateBlock(block, BlockState::NeedsRecompile);
|
InvalidateBlock(block, BlockState::NeedsRecompile);
|
||||||
CompileOrRevalidateBlock(start_pc);
|
CompileOrRevalidateBlock(start_pc);
|
||||||
|
|
||||||
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
const void* CPU::CodeCache::CreateBlockLink(Block* block, void* code, u32 newpc)
|
const void* CPU::CodeCache::CreateBlockLink(Block* block, void* code, u32 newpc)
|
||||||
|
@ -1430,7 +1447,7 @@ void CPU::CodeCache::ClearASMFunctions()
|
||||||
|
|
||||||
void CPU::CodeCache::CompileASMFunctions()
|
void CPU::CodeCache::CompileASMFunctions()
|
||||||
{
|
{
|
||||||
s_code_buffer.WriteProtect(false);
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
const u32 asm_size = EmitASMFunctions(s_code_buffer.GetFreeCodePointer(), s_code_buffer.GetFreeCodeSpace());
|
const u32 asm_size = EmitASMFunctions(s_code_buffer.GetFreeCodePointer(), s_code_buffer.GetFreeCodeSpace());
|
||||||
|
|
||||||
|
@ -1439,13 +1456,11 @@ void CPU::CodeCache::CompileASMFunctions()
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
s_code_buffer.CommitCode(asm_size);
|
s_code_buffer.CommitCode(asm_size);
|
||||||
s_code_buffer.WriteProtect(true);
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
bool CPU::CodeCache::CompileBlock(Block* block)
|
bool CPU::CodeCache::CompileBlock(Block* block)
|
||||||
{
|
{
|
||||||
s_code_buffer.WriteProtect(false);
|
|
||||||
|
|
||||||
const void* host_code = nullptr;
|
const void* host_code = nullptr;
|
||||||
u32 host_code_size = 0;
|
u32 host_code_size = 0;
|
||||||
u32 host_far_code_size = 0;
|
u32 host_far_code_size = 0;
|
||||||
|
@ -1462,8 +1477,6 @@ bool CPU::CodeCache::CompileBlock(Block* block)
|
||||||
host_code = NewRec::g_compiler->CompileBlock(block, &host_code_size, &host_far_code_size);
|
host_code = NewRec::g_compiler->CompileBlock(block, &host_code_size, &host_far_code_size);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
s_code_buffer.WriteProtect(true);
|
|
||||||
|
|
||||||
block->host_code = host_code;
|
block->host_code = host_code;
|
||||||
|
|
||||||
if (!host_code)
|
if (!host_code)
|
||||||
|
@ -1628,7 +1641,7 @@ bool CPU::CodeCache::HasPreviouslyFaultedOnPC(u32 guest_pc)
|
||||||
|
|
||||||
void CPU::CodeCache::BackpatchLoadStore(void* host_pc, const LoadstoreBackpatchInfo& info)
|
void CPU::CodeCache::BackpatchLoadStore(void* host_pc, const LoadstoreBackpatchInfo& info)
|
||||||
{
|
{
|
||||||
s_code_buffer.WriteProtect(false);
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
#ifdef ENABLE_RECOMPILER
|
#ifdef ENABLE_RECOMPILER
|
||||||
if (g_settings.cpu_execution_mode == CPUExecutionMode::Recompiler)
|
if (g_settings.cpu_execution_mode == CPUExecutionMode::Recompiler)
|
||||||
|
@ -1639,7 +1652,7 @@ void CPU::CodeCache::BackpatchLoadStore(void* host_pc, const LoadstoreBackpatchI
|
||||||
NewRec::BackpatchLoadStore(host_pc, info);
|
NewRec::BackpatchLoadStore(host_pc, info);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
s_code_buffer.WriteProtect(true);
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif // ENABLE_RECOMPILER_SUPPORT
|
#endif // ENABLE_RECOMPILER_SUPPORT
|
||||||
|
|
|
@ -1,11 +1,15 @@
|
||||||
// SPDX-FileCopyrightText: 2019-2022 Connor McLaughlin <stenzek@gmail.com>
|
// SPDX-FileCopyrightText: 2019-2023 Connor McLaughlin <stenzek@gmail.com>
|
||||||
// SPDX-License-Identifier: (GPL-3.0 OR CC-BY-NC-ND-4.0)
|
// SPDX-License-Identifier: (GPL-3.0 OR CC-BY-NC-ND-4.0)
|
||||||
|
|
||||||
#include "jit_code_buffer.h"
|
#include "jit_code_buffer.h"
|
||||||
|
|
||||||
#include "common/align.h"
|
#include "common/align.h"
|
||||||
#include "common/assert.h"
|
#include "common/assert.h"
|
||||||
#include "common/log.h"
|
#include "common/log.h"
|
||||||
|
#include "common/memmap.h"
|
||||||
|
|
||||||
#include <algorithm>
|
#include <algorithm>
|
||||||
|
|
||||||
Log_SetChannel(JitCodeBuffer);
|
Log_SetChannel(JitCodeBuffer);
|
||||||
|
|
||||||
#if defined(_WIN32)
|
#if defined(_WIN32)
|
||||||
|
@ -15,11 +19,6 @@ Log_SetChannel(JitCodeBuffer);
|
||||||
#include <sys/mman.h>
|
#include <sys/mman.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if defined(__APPLE__) && defined(__aarch64__)
|
|
||||||
// pthread_jit_write_protect_np()
|
|
||||||
#include <pthread.h>
|
|
||||||
#endif
|
|
||||||
|
|
||||||
JitCodeBuffer::JitCodeBuffer() = default;
|
JitCodeBuffer::JitCodeBuffer() = default;
|
||||||
|
|
||||||
JitCodeBuffer::JitCodeBuffer(u32 size, u32 far_code_size)
|
JitCodeBuffer::JitCodeBuffer(u32 size, u32 far_code_size)
|
||||||
|
@ -235,7 +234,7 @@ void JitCodeBuffer::CommitFarCode(u32 length)
|
||||||
|
|
||||||
void JitCodeBuffer::Reset()
|
void JitCodeBuffer::Reset()
|
||||||
{
|
{
|
||||||
WriteProtect(false);
|
MemMap::BeginCodeWrite();
|
||||||
|
|
||||||
m_free_code_ptr = m_code_ptr + m_guard_size + m_code_reserve_size;
|
m_free_code_ptr = m_code_ptr + m_guard_size + m_code_reserve_size;
|
||||||
m_code_used = 0;
|
m_code_used = 0;
|
||||||
|
@ -250,7 +249,7 @@ void JitCodeBuffer::Reset()
|
||||||
FlushInstructionCache(m_free_far_code_ptr, m_far_code_size);
|
FlushInstructionCache(m_free_far_code_ptr, m_far_code_size);
|
||||||
}
|
}
|
||||||
|
|
||||||
WriteProtect(true);
|
MemMap::EndCodeWrite();
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCodeBuffer::Align(u32 alignment, u8 padding_value)
|
void JitCodeBuffer::Align(u32 alignment, u8 padding_value)
|
||||||
|
@ -275,26 +274,3 @@ void JitCodeBuffer::FlushInstructionCache(void* address, u32 size)
|
||||||
#error Unknown platform.
|
#error Unknown platform.
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
#if defined(__APPLE__) && defined(__aarch64__)
|
|
||||||
|
|
||||||
void JitCodeBuffer::WriteProtect(bool enabled)
|
|
||||||
{
|
|
||||||
static bool initialized = false;
|
|
||||||
static bool needs_write_protect = false;
|
|
||||||
|
|
||||||
if (!initialized)
|
|
||||||
{
|
|
||||||
initialized = true;
|
|
||||||
needs_write_protect = (pthread_jit_write_protect_supported_np() != 0);
|
|
||||||
if (needs_write_protect)
|
|
||||||
Log_InfoPrint("pthread_jit_write_protect_np() will be used before writing to JIT space.");
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!needs_write_protect)
|
|
||||||
return;
|
|
||||||
|
|
||||||
pthread_jit_write_protect_np(enabled ? 1 : 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
#endif
|
|
|
@ -47,13 +47,6 @@ public:
|
||||||
/// Flushes the instruction cache on the host for the specified range.
|
/// Flushes the instruction cache on the host for the specified range.
|
||||||
static void FlushInstructionCache(void* address, u32 size);
|
static void FlushInstructionCache(void* address, u32 size);
|
||||||
|
|
||||||
/// For Apple Silicon - Toggles write protection on the JIT space.
|
|
||||||
#if defined(__APPLE__) && defined(__aarch64__)
|
|
||||||
static void WriteProtect(bool enabled);
|
|
||||||
#else
|
|
||||||
ALWAYS_INLINE static void WriteProtect(bool enabled) {}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
private:
|
private:
|
||||||
u8* m_code_ptr = nullptr;
|
u8* m_code_ptr = nullptr;
|
||||||
u8* m_free_code_ptr = nullptr;
|
u8* m_free_code_ptr = nullptr;
|
||||||
|
|
Loading…
Reference in New Issue