forked from ShuriZma/suyu
shader_jit_x64: Rewrite flow control to support arbitrary CALL and JMP instructions.
This commit is contained in:
parent
135aec7bea
commit
4632791a40
|
@ -137,6 +137,15 @@ static const u8 NO_SRC_REG_SWIZZLE = 0x1b;
|
||||||
/// Raw constant for the destination register enable mask that indicates all components are enabled
|
/// Raw constant for the destination register enable mask that indicates all components are enabled
|
||||||
static const u8 NO_DEST_REG_MASK = 0xf;
|
static const u8 NO_DEST_REG_MASK = 0xf;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the vertex shader instruction for a given offset in the current shader program
|
||||||
|
* @param offset Offset in the current shader program of the instruction
|
||||||
|
* @return Instruction at the specified offset
|
||||||
|
*/
|
||||||
|
static Instruction GetVertexShaderInstruction(size_t offset) {
|
||||||
|
return { g_state.vs.program_code[offset] };
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Loads and swizzles a source register into the specified XMM register.
|
* Loads and swizzles a source register into the specified XMM register.
|
||||||
* @param instr VS instruction, used for determining how to load the source register
|
* @param instr VS instruction, used for determining how to load the source register
|
||||||
|
@ -564,10 +573,23 @@ void JitCompiler::Compile_END(Instruction instr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_CALL(Instruction instr) {
|
void JitCompiler::Compile_CALL(Instruction instr) {
|
||||||
unsigned offset = instr.flow_control.dest_offset;
|
// Need to advance the return address past the proceeding instructions, this is the number of bytes to skip
|
||||||
while (offset < (instr.flow_control.dest_offset + instr.flow_control.num_instructions)) {
|
constexpr unsigned SKIP = 21;
|
||||||
Compile_NextInstr(&offset);
|
const uintptr_t start = reinterpret_cast<uintptr_t>(GetCodePtr());
|
||||||
}
|
|
||||||
|
// Push return address - not using CALL because we also want to push the offset of the return before jumping
|
||||||
|
MOV(64, R(RAX), ImmPtr(GetCodePtr() + SKIP));
|
||||||
|
PUSH(RAX);
|
||||||
|
|
||||||
|
// Push offset of the return
|
||||||
|
PUSH(32, Imm32(instr.flow_control.dest_offset + instr.flow_control.num_instructions));
|
||||||
|
|
||||||
|
// Jump
|
||||||
|
FixupBranch b = J(true);
|
||||||
|
fixup_branches.push_back({ b, instr.flow_control.dest_offset });
|
||||||
|
|
||||||
|
// Make sure that if the above code changes, SKIP gets updated
|
||||||
|
ASSERT(reinterpret_cast<uintptr_t>(GetCodePtr()) - start == SKIP);
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_CALLC(Instruction instr) {
|
void JitCompiler::Compile_CALLC(Instruction instr) {
|
||||||
|
@ -645,8 +667,8 @@ void JitCompiler::Compile_MAD(Instruction instr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_IF(Instruction instr) {
|
void JitCompiler::Compile_IF(Instruction instr) {
|
||||||
ASSERT_MSG(instr.flow_control.dest_offset > *offset_ptr, "Backwards if-statements (%d -> %d) not supported",
|
ASSERT_MSG(instr.flow_control.dest_offset > last_program_counter, "Backwards if-statements (%d -> %d) not supported",
|
||||||
*offset_ptr, instr.flow_control.dest_offset.Value());
|
last_program_counter, instr.flow_control.dest_offset.Value());
|
||||||
|
|
||||||
// Evaluate the "IF" condition
|
// Evaluate the "IF" condition
|
||||||
if (instr.opcode.Value() == OpCode::Id::IFU) {
|
if (instr.opcode.Value() == OpCode::Id::IFU) {
|
||||||
|
@ -677,8 +699,8 @@ void JitCompiler::Compile_IF(Instruction instr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_LOOP(Instruction instr) {
|
void JitCompiler::Compile_LOOP(Instruction instr) {
|
||||||
ASSERT_MSG(instr.flow_control.dest_offset > *offset_ptr, "Backwards loops (%d -> %d) not supported",
|
ASSERT_MSG(instr.flow_control.dest_offset > last_program_counter, "Backwards loops (%d -> %d) not supported",
|
||||||
*offset_ptr, instr.flow_control.dest_offset.Value());
|
last_program_counter, instr.flow_control.dest_offset.Value());
|
||||||
ASSERT_MSG(!looping, "Nested loops not supported");
|
ASSERT_MSG(!looping, "Nested loops not supported");
|
||||||
|
|
||||||
looping = true;
|
looping = true;
|
||||||
|
@ -706,9 +728,6 @@ void JitCompiler::Compile_LOOP(Instruction instr) {
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_JMP(Instruction instr) {
|
void JitCompiler::Compile_JMP(Instruction instr) {
|
||||||
ASSERT_MSG(instr.flow_control.dest_offset > *offset_ptr, "Backwards jumps (%d -> %d) not supported",
|
|
||||||
*offset_ptr, instr.flow_control.dest_offset.Value());
|
|
||||||
|
|
||||||
if (instr.opcode.Value() == OpCode::Id::JMPC)
|
if (instr.opcode.Value() == OpCode::Id::JMPC)
|
||||||
Compile_EvaluateCondition(instr);
|
Compile_EvaluateCondition(instr);
|
||||||
else if (instr.opcode.Value() == OpCode::Id::JMPU)
|
else if (instr.opcode.Value() == OpCode::Id::JMPU)
|
||||||
|
@ -718,31 +737,42 @@ void JitCompiler::Compile_JMP(Instruction instr) {
|
||||||
|
|
||||||
bool inverted_condition = (instr.opcode.Value() == OpCode::Id::JMPU) &&
|
bool inverted_condition = (instr.opcode.Value() == OpCode::Id::JMPU) &&
|
||||||
(instr.flow_control.num_instructions & 1);
|
(instr.flow_control.num_instructions & 1);
|
||||||
|
|
||||||
FixupBranch b = J_CC(inverted_condition ? CC_Z : CC_NZ, true);
|
FixupBranch b = J_CC(inverted_condition ? CC_Z : CC_NZ, true);
|
||||||
|
fixup_branches.push_back({ b, instr.flow_control.dest_offset });
|
||||||
Compile_Block(instr.flow_control.dest_offset);
|
|
||||||
|
|
||||||
SetJumpTarget(b);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_Block(unsigned end) {
|
void JitCompiler::Compile_Block(unsigned end) {
|
||||||
// Save current offset pointer
|
while (program_counter < end) {
|
||||||
unsigned* prev_offset_ptr = offset_ptr;
|
Compile_NextInstr();
|
||||||
unsigned offset = *prev_offset_ptr;
|
}
|
||||||
|
|
||||||
while (offset < end)
|
|
||||||
Compile_NextInstr(&offset);
|
|
||||||
|
|
||||||
// Restore current offset pointer
|
|
||||||
offset_ptr = prev_offset_ptr;
|
|
||||||
*offset_ptr = offset;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void JitCompiler::Compile_NextInstr(unsigned* offset) {
|
void JitCompiler::Compile_Return() {
|
||||||
offset_ptr = offset;
|
// Peek return offset on the stack and check if we're at that offset
|
||||||
|
MOV(64, R(RAX), MDisp(RSP, 0));
|
||||||
|
CMP(32, R(RAX), Imm32(program_counter));
|
||||||
|
|
||||||
Instruction instr;
|
// If so, jump back to before CALL
|
||||||
std::memcpy(&instr, &g_state.vs.program_code[(*offset_ptr)++], sizeof(Instruction));
|
FixupBranch b = J_CC(CC_NZ, true);
|
||||||
|
ADD(64, R(RSP), Imm32(8)); // Ignore return offset that's on the stack
|
||||||
|
POP(RAX); // Pop off return address
|
||||||
|
JMPptr(R(RAX));
|
||||||
|
SetJumpTarget(b);
|
||||||
|
}
|
||||||
|
|
||||||
|
void JitCompiler::Compile_NextInstr() {
|
||||||
|
last_program_counter = program_counter;
|
||||||
|
|
||||||
|
auto search = return_offsets.find(program_counter);
|
||||||
|
if (search != return_offsets.end()) {
|
||||||
|
Compile_Return();
|
||||||
|
}
|
||||||
|
|
||||||
|
ASSERT_MSG(code_ptr[program_counter] == nullptr, "Tried to compile already compiled shader location!");
|
||||||
|
code_ptr[program_counter] = GetCodePtr();
|
||||||
|
|
||||||
|
Instruction instr = GetVertexShaderInstruction(program_counter++);
|
||||||
|
|
||||||
OpCode::Id opcode = instr.opcode.Value();
|
OpCode::Id opcode = instr.opcode.Value();
|
||||||
auto instr_func = instr_table[static_cast<unsigned>(opcode)];
|
auto instr_func = instr_table[static_cast<unsigned>(opcode)];
|
||||||
|
@ -757,9 +787,24 @@ void JitCompiler::Compile_NextInstr(unsigned* offset) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void JitCompiler::FindReturnOffsets() {
|
||||||
|
return_offsets.clear();
|
||||||
|
|
||||||
|
for (size_t offset = 0; offset < g_state.vs.program_code.size(); ++offset) {
|
||||||
|
Instruction instr = GetVertexShaderInstruction(offset);
|
||||||
|
|
||||||
|
switch (instr.opcode.Value()) {
|
||||||
|
case OpCode::Id::CALL:
|
||||||
|
case OpCode::Id::CALLC:
|
||||||
|
case OpCode::Id::CALLU:
|
||||||
|
return_offsets.insert(instr.flow_control.dest_offset + instr.flow_control.num_instructions);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
CompiledShader* JitCompiler::Compile() {
|
CompiledShader* JitCompiler::Compile() {
|
||||||
const u8* start = GetCodePtr();
|
const u8* start = GetCodePtr();
|
||||||
unsigned offset = g_state.regs.vs.main_offset;
|
|
||||||
|
|
||||||
// The stack pointer is 8 modulo 16 at the entry of a procedure
|
// The stack pointer is 8 modulo 16 at the entry of a procedure
|
||||||
ABI_PushRegistersAndAdjustStack(ABI_ALL_CALLEE_SAVED, 8);
|
ABI_PushRegistersAndAdjustStack(ABI_ALL_CALLEE_SAVED, 8);
|
||||||
|
@ -782,10 +827,27 @@ CompiledShader* JitCompiler::Compile() {
|
||||||
MOV(PTRBITS, R(RAX), ImmPtr(&neg));
|
MOV(PTRBITS, R(RAX), ImmPtr(&neg));
|
||||||
MOVAPS(NEGBIT, MatR(RAX));
|
MOVAPS(NEGBIT, MatR(RAX));
|
||||||
|
|
||||||
looping = false;
|
// Find all `CALL` instructions and identify return locations
|
||||||
|
FindReturnOffsets();
|
||||||
|
|
||||||
while (offset < g_state.vs.program_code.size()) {
|
// Reset flow control state
|
||||||
Compile_NextInstr(&offset);
|
last_program_counter = 0;
|
||||||
|
program_counter = 0;
|
||||||
|
looping = false;
|
||||||
|
code_ptr.fill(nullptr);
|
||||||
|
fixup_branches.clear();
|
||||||
|
|
||||||
|
// Jump to start of the shader program
|
||||||
|
if (g_state.regs.vs.main_offset != 0) {
|
||||||
|
fixup_branches.push_back({ J(true), g_state.regs.vs.main_offset });
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compile entire program
|
||||||
|
Compile_Block(static_cast<unsigned>(g_state.vs.program_code.size()));
|
||||||
|
|
||||||
|
// Set the target for any incomplete branches now that the entire shader program has been emitted
|
||||||
|
for (const auto& branch : fixup_branches) {
|
||||||
|
SetJumpTarget(branch.first, code_ptr[branch.second]);
|
||||||
}
|
}
|
||||||
|
|
||||||
return (CompiledShader*)start;
|
return (CompiledShader*)start;
|
||||||
|
|
|
@ -4,6 +4,9 @@
|
||||||
|
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
|
#include <set>
|
||||||
|
#include <utility>
|
||||||
|
|
||||||
#include <nihstro/shader_bytecode.h>
|
#include <nihstro/shader_bytecode.h>
|
||||||
|
|
||||||
#include "common/x64/emitter.h"
|
#include "common/x64/emitter.h"
|
||||||
|
@ -66,8 +69,9 @@ public:
|
||||||
void Compile_MAD(Instruction instr);
|
void Compile_MAD(Instruction instr);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
|
||||||
void Compile_Block(unsigned end);
|
void Compile_Block(unsigned end);
|
||||||
void Compile_NextInstr(unsigned* offset);
|
void Compile_NextInstr();
|
||||||
|
|
||||||
void Compile_SwizzleSrc(Instruction instr, unsigned src_num, SourceRegister src_reg, Gen::X64Reg dest);
|
void Compile_SwizzleSrc(Instruction instr, unsigned src_num, SourceRegister src_reg, Gen::X64Reg dest);
|
||||||
void Compile_DestEnable(Instruction instr, Gen::X64Reg dest);
|
void Compile_DestEnable(Instruction instr, Gen::X64Reg dest);
|
||||||
|
@ -81,13 +85,31 @@ private:
|
||||||
void Compile_EvaluateCondition(Instruction instr);
|
void Compile_EvaluateCondition(Instruction instr);
|
||||||
void Compile_UniformCondition(Instruction instr);
|
void Compile_UniformCondition(Instruction instr);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Emits the code to conditionally return from a subroutine envoked by the `CALL` instruction.
|
||||||
|
*/
|
||||||
|
void Compile_Return();
|
||||||
|
|
||||||
BitSet32 PersistentCallerSavedRegs();
|
BitSet32 PersistentCallerSavedRegs();
|
||||||
|
|
||||||
/// Pointer to the variable that stores the current Pica code offset. Used to handle nested code blocks.
|
/**
|
||||||
unsigned* offset_ptr = nullptr;
|
* Analyzes the entire shader program for `CALL` instructions before emitting any code,
|
||||||
|
* identifying the locations where a return needs to be inserted.
|
||||||
|
*/
|
||||||
|
void FindReturnOffsets();
|
||||||
|
|
||||||
/// Set to true if currently in a loop, used to check for the existence of nested loops
|
/// Mapping of Pica VS instructions to pointers in the emitted code
|
||||||
bool looping = false;
|
std::array<const u8*, 1024> code_ptr;
|
||||||
|
|
||||||
|
/// Offsets in code where a return needs to be inserted
|
||||||
|
std::set<unsigned> return_offsets;
|
||||||
|
|
||||||
|
unsigned last_program_counter; ///< Offset of the most recent instruction decoded
|
||||||
|
unsigned program_counter; ///< Offset of the next instruction to decode
|
||||||
|
bool looping = false; ///< True if compiling a loop, used to check for nested loops
|
||||||
|
|
||||||
|
/// Branches that need to be fixed up once the entire shader program is compiled
|
||||||
|
std::vector<std::pair<Gen::FixupBranch, unsigned>> fixup_branches;
|
||||||
};
|
};
|
||||||
|
|
||||||
} // Shader
|
} // Shader
|
||||||
|
|
Loading…
Reference in New Issue