2015-05-24 04:55:12 +00:00
|
|
|
// Copyright 2008 Dolphin Emulator Project
|
2015-05-17 23:08:10 +00:00
|
|
|
// Licensed under GPLv2+
|
2013-04-18 03:09:55 +00:00
|
|
|
// Refer to the license.txt file included.
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2014-02-19 01:27:20 +00:00
|
|
|
#include "Common/Atomic.h"
|
|
|
|
#include "Common/ChunkFile.h"
|
2014-09-08 01:06:58 +00:00
|
|
|
#include "Common/CommonTypes.h"
|
2014-02-18 11:18:47 +00:00
|
|
|
#include "Common/MathUtil.h"
|
|
|
|
#include "Common/Thread.h"
|
|
|
|
#include "Core/ConfigManager.h"
|
|
|
|
#include "Core/Core.h"
|
|
|
|
#include "Core/CoreTiming.h"
|
|
|
|
#include "Core/HW/GPFifo.h"
|
|
|
|
#include "Core/HW/Memmap.h"
|
|
|
|
#include "Core/HW/MMIO.h"
|
|
|
|
#include "Core/HW/ProcessorInterface.h"
|
|
|
|
#include "Core/HW/SystemTimers.h"
|
|
|
|
#include "VideoCommon/CommandProcessor.h"
|
|
|
|
#include "VideoCommon/Fifo.h"
|
|
|
|
#include "VideoCommon/PixelEngine.h"
|
|
|
|
#include "VideoCommon/VideoCommon.h"
|
|
|
|
#include "VideoCommon/VideoConfig.h"
|
2010-06-09 01:37:08 +00:00
|
|
|
|
|
|
|
namespace CommandProcessor
|
|
|
|
{
|
|
|
|
|
2014-07-08 13:58:25 +00:00
|
|
|
static int et_UpdateInterrupts;
|
2010-06-09 01:37:08 +00:00
|
|
|
|
|
|
|
// TODO(ector): Warn on bbox read/write
|
|
|
|
|
|
|
|
// STATE_TO_SAVE
|
|
|
|
SCPFifoStruct fifo;
|
2014-07-08 13:58:25 +00:00
|
|
|
static UCPStatusReg m_CPStatusReg;
|
|
|
|
static UCPCtrlReg m_CPCtrlReg;
|
|
|
|
static UCPClearReg m_CPClearReg;
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2014-07-08 13:58:25 +00:00
|
|
|
static u16 m_bboxleft;
|
|
|
|
static u16 m_bboxtop;
|
|
|
|
static u16 m_bboxright;
|
|
|
|
static u16 m_bboxbottom;
|
|
|
|
static u16 m_tokenReg;
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2010-12-11 12:42:55 +00:00
|
|
|
volatile bool interruptSet= false;
|
|
|
|
volatile bool interruptWaiting= false;
|
2010-12-13 07:56:54 +00:00
|
|
|
volatile bool interruptTokenWaiting = false;
|
|
|
|
volatile bool interruptFinishWaiting = false;
|
2010-06-24 13:28:54 +00:00
|
|
|
|
2015-03-13 22:36:31 +00:00
|
|
|
Common::Flag s_gpuMaySleep;
|
|
|
|
|
2013-02-16 01:51:09 +00:00
|
|
|
volatile u32 VITicks = CommandProcessor::m_cpClockOrigin;
|
|
|
|
|
2014-07-08 12:29:26 +00:00
|
|
|
static bool IsOnThread()
|
2011-02-08 10:37:47 +00:00
|
|
|
{
|
|
|
|
return SConfig::GetInstance().m_LocalCoreStartupParameter.bCPUThread;
|
|
|
|
}
|
|
|
|
|
2014-07-08 12:29:26 +00:00
|
|
|
static void UpdateInterrupts_Wrapper(u64 userdata, int cyclesLate)
|
2011-02-08 10:37:47 +00:00
|
|
|
{
|
|
|
|
UpdateInterrupts(userdata);
|
|
|
|
}
|
|
|
|
|
2010-06-09 01:37:08 +00:00
|
|
|
void DoState(PointerWrap &p)
|
|
|
|
{
|
2013-04-09 23:57:39 +00:00
|
|
|
p.DoPOD(m_CPStatusReg);
|
|
|
|
p.DoPOD(m_CPCtrlReg);
|
|
|
|
p.DoPOD(m_CPClearReg);
|
2010-06-09 01:37:08 +00:00
|
|
|
p.Do(m_bboxleft);
|
|
|
|
p.Do(m_bboxtop);
|
|
|
|
p.Do(m_bboxright);
|
|
|
|
p.Do(m_bboxbottom);
|
|
|
|
p.Do(m_tokenReg);
|
|
|
|
p.Do(fifo);
|
2011-02-08 10:37:47 +00:00
|
|
|
|
|
|
|
p.Do(interruptSet);
|
|
|
|
p.Do(interruptWaiting);
|
|
|
|
p.Do(interruptTokenWaiting);
|
|
|
|
p.Do(interruptFinishWaiting);
|
2010-06-09 01:37:08 +00:00
|
|
|
}
|
|
|
|
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
static inline void WriteLow(volatile u32& _reg, u16 lowbits)
|
2014-08-25 19:09:26 +00:00
|
|
|
{
|
|
|
|
Common::AtomicStore(_reg, (_reg & 0xFFFF0000) | lowbits);
|
|
|
|
}
|
|
|
|
static inline void WriteHigh(volatile u32& _reg, u16 highbits)
|
|
|
|
{
|
|
|
|
Common::AtomicStore(_reg, (_reg & 0x0000FFFF) | ((u32)highbits << 16));
|
|
|
|
}
|
|
|
|
static inline u16 ReadLow(u32 _reg)
|
|
|
|
{
|
|
|
|
return (u16)(_reg & 0xFFFF);
|
|
|
|
}
|
|
|
|
static inline u16 ReadHigh(u32 _reg)
|
|
|
|
{
|
|
|
|
return (u16)(_reg >> 16);
|
|
|
|
}
|
2010-06-09 01:37:08 +00:00
|
|
|
|
|
|
|
void Init()
|
|
|
|
{
|
|
|
|
m_CPStatusReg.Hex = 0;
|
2010-06-14 21:55:40 +00:00
|
|
|
m_CPStatusReg.CommandIdle = 1;
|
2010-06-09 01:37:08 +00:00
|
|
|
m_CPStatusReg.ReadIdle = 1;
|
|
|
|
|
|
|
|
m_CPCtrlReg.Hex = 0;
|
|
|
|
|
2012-01-02 10:20:22 +00:00
|
|
|
m_CPClearReg.Hex = 0;
|
|
|
|
|
2010-06-09 01:37:08 +00:00
|
|
|
m_bboxleft = 0;
|
|
|
|
m_bboxtop = 0;
|
|
|
|
m_bboxright = 640;
|
|
|
|
m_bboxbottom = 480;
|
|
|
|
|
|
|
|
m_tokenReg = 0;
|
2013-10-29 05:23:17 +00:00
|
|
|
|
2010-06-09 01:37:08 +00:00
|
|
|
memset(&fifo,0,sizeof(fifo));
|
2010-06-14 21:55:40 +00:00
|
|
|
fifo.bFF_Breakpoint = 0;
|
2013-04-24 13:21:54 +00:00
|
|
|
fifo.bFF_HiWatermark = 0;
|
2010-12-11 12:42:55 +00:00
|
|
|
fifo.bFF_HiWatermarkInt = 0;
|
2013-04-24 13:21:54 +00:00
|
|
|
fifo.bFF_LoWatermark = 0;
|
2010-12-11 12:42:55 +00:00
|
|
|
fifo.bFF_LoWatermarkInt = 0;
|
|
|
|
|
|
|
|
interruptSet = false;
|
2013-04-24 13:21:54 +00:00
|
|
|
interruptWaiting = false;
|
2010-12-13 07:56:54 +00:00
|
|
|
interruptFinishWaiting = false;
|
|
|
|
interruptTokenWaiting = false;
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2013-04-24 13:21:54 +00:00
|
|
|
et_UpdateInterrupts = CoreTiming::RegisterEvent("CPInterrupt", UpdateInterrupts_Wrapper);
|
2010-06-09 01:37:08 +00:00
|
|
|
}
|
|
|
|
|
2014-02-02 13:16:43 +00:00
|
|
|
void RegisterMMIO(MMIO::Mapping* mmio, u32 base)
|
2010-06-09 01:37:08 +00:00
|
|
|
{
|
2014-02-02 13:16:43 +00:00
|
|
|
struct {
|
|
|
|
u32 addr;
|
|
|
|
u16* ptr;
|
|
|
|
bool readonly;
|
|
|
|
bool writes_align_to_32_bytes;
|
|
|
|
} directly_mapped_vars[] = {
|
|
|
|
{ FIFO_TOKEN_REGISTER, &m_tokenReg },
|
|
|
|
|
|
|
|
// Bounding box registers are read only.
|
|
|
|
{ FIFO_BOUNDING_BOX_LEFT, &m_bboxleft, true },
|
|
|
|
{ FIFO_BOUNDING_BOX_RIGHT, &m_bboxright, true },
|
|
|
|
{ FIFO_BOUNDING_BOX_TOP, &m_bboxtop, true },
|
|
|
|
{ FIFO_BOUNDING_BOX_BOTTOM, &m_bboxbottom, true },
|
|
|
|
|
|
|
|
// Some FIFO addresses need to be aligned on 32 bytes on write - only
|
|
|
|
// the high part can be written directly without a mask.
|
|
|
|
{ FIFO_BASE_LO, MMIO::Utils::LowPart(&fifo.CPBase), false, true },
|
|
|
|
{ FIFO_BASE_HI, MMIO::Utils::HighPart(&fifo.CPBase) },
|
|
|
|
{ FIFO_END_LO, MMIO::Utils::LowPart(&fifo.CPEnd), false, true },
|
|
|
|
{ FIFO_END_HI, MMIO::Utils::HighPart(&fifo.CPEnd) },
|
|
|
|
{ FIFO_HI_WATERMARK_LO, MMIO::Utils::LowPart(&fifo.CPHiWatermark) },
|
|
|
|
{ FIFO_HI_WATERMARK_HI, MMIO::Utils::HighPart(&fifo.CPHiWatermark) },
|
|
|
|
{ FIFO_LO_WATERMARK_LO, MMIO::Utils::LowPart(&fifo.CPLoWatermark) },
|
|
|
|
{ FIFO_LO_WATERMARK_HI, MMIO::Utils::HighPart(&fifo.CPLoWatermark) },
|
|
|
|
// FIFO_RW_DISTANCE has some complex read code different for
|
|
|
|
// single/dual core.
|
|
|
|
{ FIFO_WRITE_POINTER_LO, MMIO::Utils::LowPart(&fifo.CPWritePointer), false, true },
|
|
|
|
{ FIFO_WRITE_POINTER_HI, MMIO::Utils::HighPart(&fifo.CPWritePointer) },
|
|
|
|
// FIFO_READ_POINTER has different code for single/dual core.
|
|
|
|
};
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
|
2014-02-02 13:16:43 +00:00
|
|
|
for (auto& mapped_var : directly_mapped_vars)
|
2010-06-09 01:37:08 +00:00
|
|
|
{
|
2014-02-02 13:16:43 +00:00
|
|
|
u16 wmask = mapped_var.writes_align_to_32_bytes ? 0xFFE0 : 0xFFFF;
|
|
|
|
mmio->Register(base | mapped_var.addr,
|
|
|
|
MMIO::DirectRead<u16>(mapped_var.ptr),
|
|
|
|
mapped_var.readonly
|
|
|
|
? MMIO::InvalidWrite<u16>()
|
|
|
|
: MMIO::DirectWrite<u16>(mapped_var.ptr, wmask)
|
|
|
|
);
|
2010-06-09 01:37:08 +00:00
|
|
|
}
|
2010-06-15 14:24:01 +00:00
|
|
|
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
mmio->Register(base | FIFO_BP_LO,
|
|
|
|
MMIO::DirectRead<u16>(MMIO::Utils::LowPart(&fifo.CPBreakpoint)),
|
|
|
|
MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
WriteLow(fifo.CPBreakpoint, val & 0xffe0);
|
|
|
|
})
|
|
|
|
);
|
|
|
|
mmio->Register(base | FIFO_BP_HI,
|
|
|
|
MMIO::DirectRead<u16>(MMIO::Utils::HighPart(&fifo.CPBreakpoint)),
|
|
|
|
MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
WriteHigh(fifo.CPBreakpoint, val);
|
|
|
|
})
|
|
|
|
);
|
|
|
|
|
2014-02-02 13:16:43 +00:00
|
|
|
// Timing and metrics MMIOs are stubbed with fixed values.
|
|
|
|
struct {
|
|
|
|
u32 addr;
|
|
|
|
u16 value;
|
|
|
|
} metrics_mmios[] = {
|
|
|
|
{ XF_RASBUSY_L, 0 },
|
|
|
|
{ XF_RASBUSY_H, 0 },
|
|
|
|
{ XF_CLKS_L, 0 },
|
|
|
|
{ XF_CLKS_H, 0 },
|
|
|
|
{ XF_WAIT_IN_L, 0 },
|
|
|
|
{ XF_WAIT_IN_H, 0 },
|
|
|
|
{ XF_WAIT_OUT_L, 0 },
|
|
|
|
{ XF_WAIT_OUT_H, 0 },
|
|
|
|
{ VCACHE_METRIC_CHECK_L, 0 },
|
|
|
|
{ VCACHE_METRIC_CHECK_H, 0 },
|
|
|
|
{ VCACHE_METRIC_MISS_L, 0 },
|
|
|
|
{ VCACHE_METRIC_MISS_H, 0 },
|
|
|
|
{ VCACHE_METRIC_STALL_L, 0 },
|
|
|
|
{ VCACHE_METRIC_STALL_H, 0 },
|
|
|
|
{ CLKS_PER_VTX_OUT, 4 },
|
|
|
|
};
|
|
|
|
for (auto& metrics_mmio : metrics_mmios)
|
2010-06-09 01:37:08 +00:00
|
|
|
{
|
2014-02-02 13:16:43 +00:00
|
|
|
mmio->Register(base | metrics_mmio.addr,
|
|
|
|
MMIO::Constant<u16>(metrics_mmio.value),
|
|
|
|
MMIO::InvalidWrite<u16>()
|
|
|
|
);
|
|
|
|
}
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2014-02-02 13:16:43 +00:00
|
|
|
mmio->Register(base | STATUS_REGISTER,
|
|
|
|
MMIO::ComplexRead<u16>([](u32) {
|
|
|
|
SetCpStatusRegister();
|
|
|
|
return m_CPStatusReg.Hex;
|
|
|
|
}),
|
|
|
|
MMIO::InvalidWrite<u16>()
|
|
|
|
);
|
|
|
|
|
|
|
|
mmio->Register(base | CTRL_REGISTER,
|
|
|
|
MMIO::DirectRead<u16>(&m_CPCtrlReg.Hex),
|
|
|
|
MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
UCPCtrlReg tmp(val);
|
|
|
|
m_CPCtrlReg.Hex = tmp.Hex;
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
SetCpControlRegister();
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
RunGpu();
|
2014-02-02 13:16:43 +00:00
|
|
|
})
|
|
|
|
);
|
|
|
|
|
|
|
|
mmio->Register(base | CLEAR_REGISTER,
|
|
|
|
MMIO::DirectRead<u16>(&m_CPClearReg.Hex),
|
|
|
|
MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
UCPClearReg tmp(val);
|
|
|
|
m_CPClearReg.Hex = tmp.Hex;
|
2012-01-21 13:58:29 +00:00
|
|
|
SetCpClearRegister();
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
RunGpu();
|
2014-02-02 13:16:43 +00:00
|
|
|
})
|
|
|
|
);
|
|
|
|
|
|
|
|
mmio->Register(base | PERF_SELECT,
|
|
|
|
MMIO::InvalidRead<u16>(),
|
|
|
|
MMIO::Nop<u16>()
|
|
|
|
);
|
|
|
|
|
|
|
|
// Some MMIOs have different handlers for single core vs. dual core mode.
|
|
|
|
mmio->Register(base | FIFO_RW_DISTANCE_LO,
|
|
|
|
IsOnThread()
|
|
|
|
? MMIO::ComplexRead<u16>([](u32) {
|
|
|
|
if (fifo.CPWritePointer >= fifo.SafeCPReadPointer)
|
|
|
|
return ReadLow(fifo.CPWritePointer - fifo.SafeCPReadPointer);
|
|
|
|
else
|
|
|
|
return ReadLow(fifo.CPEnd - fifo.SafeCPReadPointer + fifo.CPWritePointer - fifo.CPBase + 32);
|
|
|
|
})
|
|
|
|
: MMIO::DirectRead<u16>(MMIO::Utils::LowPart(&fifo.CPReadWriteDistance)),
|
|
|
|
MMIO::DirectWrite<u16>(MMIO::Utils::LowPart(&fifo.CPReadWriteDistance), 0xFFE0)
|
|
|
|
);
|
|
|
|
mmio->Register(base | FIFO_RW_DISTANCE_HI,
|
|
|
|
IsOnThread()
|
|
|
|
? MMIO::ComplexRead<u16>([](u32) {
|
|
|
|
if (fifo.CPWritePointer >= fifo.SafeCPReadPointer)
|
|
|
|
return ReadHigh(fifo.CPWritePointer - fifo.SafeCPReadPointer);
|
|
|
|
else
|
|
|
|
return ReadHigh(fifo.CPEnd - fifo.SafeCPReadPointer + fifo.CPWritePointer - fifo.CPBase + 32);
|
|
|
|
})
|
|
|
|
: MMIO::DirectRead<u16>(MMIO::Utils::HighPart(&fifo.CPReadWriteDistance)),
|
|
|
|
MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
WriteHigh(fifo.CPReadWriteDistance, val);
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
SyncGPU(SYNC_GPU_OTHER);
|
2014-02-02 13:16:43 +00:00
|
|
|
if (fifo.CPReadWriteDistance == 0)
|
|
|
|
{
|
|
|
|
GPFifo::ResetGatherPipe();
|
|
|
|
ResetVideoBuffer();
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
ResetVideoBuffer();
|
|
|
|
}
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
RunGpu();
|
2014-02-02 13:16:43 +00:00
|
|
|
})
|
|
|
|
);
|
|
|
|
mmio->Register(base | FIFO_READ_POINTER_LO,
|
|
|
|
IsOnThread()
|
|
|
|
? MMIO::DirectRead<u16>(MMIO::Utils::LowPart(&fifo.SafeCPReadPointer))
|
|
|
|
: MMIO::DirectRead<u16>(MMIO::Utils::LowPart(&fifo.CPReadPointer)),
|
|
|
|
MMIO::DirectWrite<u16>(MMIO::Utils::LowPart(&fifo.CPReadPointer), 0xFFE0)
|
|
|
|
);
|
|
|
|
mmio->Register(base | FIFO_READ_POINTER_HI,
|
|
|
|
IsOnThread()
|
|
|
|
? MMIO::DirectRead<u16>(MMIO::Utils::HighPart(&fifo.SafeCPReadPointer))
|
|
|
|
: MMIO::DirectRead<u16>(MMIO::Utils::HighPart(&fifo.CPReadPointer)),
|
|
|
|
IsOnThread()
|
|
|
|
? MMIO::ComplexWrite<u16>([](u32, u16 val) {
|
|
|
|
WriteHigh(fifo.CPReadPointer, val);
|
|
|
|
fifo.SafeCPReadPointer = fifo.CPReadPointer;
|
|
|
|
})
|
|
|
|
: MMIO::DirectWrite<u16>(MMIO::Utils::HighPart(&fifo.CPReadPointer))
|
|
|
|
);
|
|
|
|
}
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2014-09-30 05:22:57 +00:00
|
|
|
void GatherPipeBursted()
|
2010-06-09 01:37:08 +00:00
|
|
|
{
|
2014-11-14 06:07:11 +00:00
|
|
|
if (IsOnThread())
|
|
|
|
SetCPStatusFromCPU();
|
|
|
|
|
2012-03-19 01:54:58 +00:00
|
|
|
ProcessFifoEvents();
|
2010-06-09 01:37:08 +00:00
|
|
|
// if we aren't linked, we don't care about gather pipe data
|
2010-06-14 21:55:40 +00:00
|
|
|
if (!m_CPCtrlReg.GPLinkEnable)
|
2010-06-15 14:24:01 +00:00
|
|
|
{
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
if (IsOnThread() && !g_use_deterministic_gpu_thread)
|
2012-03-19 01:54:58 +00:00
|
|
|
{
|
2013-03-31 23:10:21 +00:00
|
|
|
// In multibuffer mode is not allowed write in the same FIFO attached to the GPU.
|
2012-03-19 01:54:58 +00:00
|
|
|
// Fix Pokemon XD in DC mode.
|
2014-03-10 11:30:55 +00:00
|
|
|
if ((ProcessorInterface::Fifo_CPUEnd == fifo.CPEnd) &&
|
|
|
|
(ProcessorInterface::Fifo_CPUBase == fifo.CPBase) &&
|
|
|
|
fifo.CPReadWriteDistance > 0)
|
2012-03-19 01:54:58 +00:00
|
|
|
{
|
2015-03-05 20:14:46 +00:00
|
|
|
FlushGpu();
|
2012-03-19 01:54:58 +00:00
|
|
|
}
|
|
|
|
}
|
2015-03-05 16:12:24 +00:00
|
|
|
RunGpu();
|
2010-06-09 01:37:08 +00:00
|
|
|
return;
|
2010-06-15 14:24:01 +00:00
|
|
|
}
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2013-03-31 23:10:21 +00:00
|
|
|
// update the fifo pointer
|
2014-11-17 05:05:16 +00:00
|
|
|
if (fifo.CPWritePointer == fifo.CPEnd)
|
2010-06-15 14:24:01 +00:00
|
|
|
fifo.CPWritePointer = fifo.CPBase;
|
2010-06-09 01:37:08 +00:00
|
|
|
else
|
2010-06-15 14:24:01 +00:00
|
|
|
fifo.CPWritePointer += GATHER_PIPE_SIZE;
|
2010-06-09 01:37:08 +00:00
|
|
|
|
2014-09-22 06:49:09 +00:00
|
|
|
if (m_CPCtrlReg.GPReadEnable && m_CPCtrlReg.GPLinkEnable)
|
|
|
|
{
|
|
|
|
ProcessorInterface::Fifo_CPUWritePointer = fifo.CPWritePointer;
|
|
|
|
ProcessorInterface::Fifo_CPUBase = fifo.CPBase;
|
|
|
|
ProcessorInterface::Fifo_CPUEnd = fifo.CPEnd;
|
|
|
|
}
|
|
|
|
|
2014-11-17 05:05:16 +00:00
|
|
|
// If the game is running close to overflowing, make the exception checking more frequent.
|
|
|
|
if (fifo.bFF_HiWatermark)
|
|
|
|
CoreTiming::ForceExceptionCheck(0);
|
|
|
|
|
2010-06-15 14:24:01 +00:00
|
|
|
Common::AtomicAdd(fifo.CPReadWriteDistance, GATHER_PIPE_SIZE);
|
2010-06-14 21:55:40 +00:00
|
|
|
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
RunGpu();
|
2010-06-15 14:24:01 +00:00
|
|
|
|
2013-04-24 13:21:54 +00:00
|
|
|
_assert_msg_(COMMANDPROCESSOR, fifo.CPReadWriteDistance <= fifo.CPEnd - fifo.CPBase,
|
2013-03-31 23:10:21 +00:00
|
|
|
"FIFO is overflowed by GatherPipe !\nCPU thread is too fast!");
|
2010-06-16 07:22:47 +00:00
|
|
|
|
2010-06-15 14:24:01 +00:00
|
|
|
// check if we are in sync
|
2014-02-16 20:30:18 +00:00
|
|
|
_assert_msg_(COMMANDPROCESSOR, fifo.CPWritePointer == ProcessorInterface::Fifo_CPUWritePointer, "FIFOs linked but out of sync");
|
|
|
|
_assert_msg_(COMMANDPROCESSOR, fifo.CPBase == ProcessorInterface::Fifo_CPUBase, "FIFOs linked but out of sync");
|
|
|
|
_assert_msg_(COMMANDPROCESSOR, fifo.CPEnd == ProcessorInterface::Fifo_CPUEnd, "FIFOs linked but out of sync");
|
2010-06-09 01:37:08 +00:00
|
|
|
}
|
|
|
|
|
2010-12-11 12:42:55 +00:00
|
|
|
void UpdateInterrupts(u64 userdata)
|
|
|
|
{
|
2013-03-20 01:51:12 +00:00
|
|
|
if (userdata)
|
2010-12-11 12:42:55 +00:00
|
|
|
{
|
|
|
|
interruptSet = true;
|
2013-03-20 01:51:12 +00:00
|
|
|
INFO_LOG(COMMANDPROCESSOR,"Interrupt set");
|
2013-04-24 13:21:54 +00:00
|
|
|
ProcessorInterface::SetInterrupt(INT_CAUSE_CP, true);
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2010-12-13 04:44:33 +00:00
|
|
|
interruptSet = false;
|
|
|
|
INFO_LOG(COMMANDPROCESSOR,"Interrupt cleared");
|
2013-04-24 13:21:54 +00:00
|
|
|
ProcessorInterface::SetInterrupt(INT_CAUSE_CP, false);
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
2014-11-14 06:07:11 +00:00
|
|
|
CoreTiming::ForceExceptionCheck(0);
|
2013-03-20 01:51:12 +00:00
|
|
|
interruptWaiting = false;
|
2015-03-05 16:12:24 +00:00
|
|
|
RunGpu();
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
|
|
|
|
2011-02-14 02:18:03 +00:00
|
|
|
void UpdateInterruptsFromVideoBackend(u64 userdata)
|
2010-06-09 01:37:08 +00:00
|
|
|
{
|
Add the 'desynced GPU thread' mode.
It's a relatively big commit (less big with -w), but it's hard to test
any of this separately...
The basic problem is that in netplay or movies, the state of the CPU
must be deterministic, including when the game receives notification
that the GPU has processed FIFO data. Dual core mode notifies the game
whenever the GPU thread actually gets around to doing the work, so it
isn't deterministic. Single core mode is because it notifies the game
'instantly' (after processing the data synchronously), but it's too slow
for many systems and games.
My old dc-netplay branch worked as follows: everything worked as normal
except the state of the CP registers was a lie, and the CPU thread only
delivered results when idle detection triggered (waiting for the GPU if
they weren't ready at that point). Usually, a game is idle iff all the
work for the frame has been done, except for a small amount of work
depending on the GPU result, so neither the CPU or the GPU waiting on
the other affected performance much. However, it's possible that the
game could be waiting for some earlier interrupt, and any of several
games which, for whatever reason, never went into a detectable idle
(even when I tried to improve the detection) would never receive results
at all. (The current method should have better compatibility, but it
also has slightly higher overhead and breaks some other things, so I
want to reimplement this, hopefully with less impact on the code, in the
future.)
With this commit, the basic idea is that the CPU thread acts as if the
work has been done instantly, like single core mode, but actually hands
it off asynchronously to the GPU thread (after backing up some data that
the game might change in memory before it's actually done). Since the
work isn't done, any feedback from the GPU to the CPU, such as real
XFB/EFB copies (virtual are OK), EFB pokes, performance queries, etc. is
broken; but most games work with these options disabled, and there is no
need to try to detect what the CPU thread is doing.
Technically: when the flag g_use_deterministic_gpu_thread (currently
stuck on) is on, the CPU thread calls RunGpu like in single core mode.
This function synchronously copies the data from the FIFO to the
internal video buffer and updates the CP registers, interrupts, etc.
However, instead of the regular ReadDataFromFifo followed by running the
opcode decoder, it runs ReadDataFromFifoOnCPU ->
OpcodeDecoder_Preprocess, which relatively quickly scans through the
FIFO data, detects SetFinish calls etc., which are immediately fired,
and saves certain associated data from memory (e.g. display lists) in
AuxBuffers (a parallel stream to the main FIFO, which is a bit slow at
the moment), before handing the data off to the GPU thread to actually
render. That makes up the bulk of this commit.
In various circumstances, including the aforementioned EFB pokes and
performance queries as well as swap requests (i.e. the end of a frame -
we don't want the CPU potentially pumping out frames too quickly and the
GPU falling behind*), SyncGPU is called to wait for actual completion.
The overhead mainly comes from OpcodeDecoder_Preprocess (which is,
again, synchronous), as well as the actual copying.
Currently, display lists and such are escrowed from main memory even
though they usually won't change over the course of a frame, and
textures are not even though they might, resulting in a small chance of
graphical glitches. When the texture locking (i.e. fault on write) code
lands, I can make this all correct and maybe a little faster.
* This suggests an alternate determinism method of just delaying results
until a short time before the end of each frame. For all I know this
might mostly work - I haven't tried it - but if any significant work
hinges on the competion of render to texture etc., the frame will be
missed.
2014-08-28 02:56:19 +00:00
|
|
|
if (!g_use_deterministic_gpu_thread)
|
|
|
|
CoreTiming::ScheduleEvent_Threadsafe(0, et_UpdateInterrupts, userdata);
|
2010-06-09 01:37:08 +00:00
|
|
|
}
|
|
|
|
|
2014-08-24 20:27:32 +00:00
|
|
|
void SetCPStatusFromGPU()
|
2010-12-11 12:42:55 +00:00
|
|
|
{
|
2013-04-24 13:21:54 +00:00
|
|
|
// breakpoint
|
2014-08-24 20:27:32 +00:00
|
|
|
if (fifo.bFF_BPEnable)
|
2013-02-16 01:51:09 +00:00
|
|
|
{
|
2014-08-24 20:27:32 +00:00
|
|
|
if (fifo.CPBreakpoint == fifo.CPReadPointer)
|
2013-02-16 01:51:09 +00:00
|
|
|
{
|
2014-08-24 20:27:32 +00:00
|
|
|
if (!fifo.bFF_Breakpoint)
|
2013-02-16 01:51:09 +00:00
|
|
|
{
|
2014-08-24 20:27:32 +00:00
|
|
|
INFO_LOG(COMMANDPROCESSOR, "Hit breakpoint at %i", fifo.CPReadPointer);
|
|
|
|
fifo.bFF_Breakpoint = true;
|
2013-02-16 01:51:09 +00:00
|
|
|
}
|
|
|
|
}
|
2010-12-11 12:42:55 +00:00
|
|
|
else
|
|
|
|
{
|
|
|
|
if (fifo.bFF_Breakpoint)
|
2010-12-12 15:25:03 +00:00
|
|
|
INFO_LOG(COMMANDPROCESSOR, "Cleared breakpoint at %i", fifo.CPReadPointer);
|
2013-02-16 01:51:09 +00:00
|
|
|
fifo.bFF_Breakpoint = false;
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
2013-02-16 01:51:09 +00:00
|
|
|
}
|
2014-08-24 20:27:32 +00:00
|
|
|
else
|
|
|
|
{
|
|
|
|
if (fifo.bFF_Breakpoint)
|
|
|
|
INFO_LOG(COMMANDPROCESSOR, "Cleared breakpoint at %i", fifo.CPReadPointer);
|
|
|
|
fifo.bFF_Breakpoint = false;
|
|
|
|
}
|
2014-11-14 06:07:11 +00:00
|
|
|
|
2014-11-14 00:46:02 +00:00
|
|
|
// overflow & underflow check
|
|
|
|
fifo.bFF_HiWatermark = (fifo.CPReadWriteDistance > fifo.CPHiWatermark);
|
|
|
|
fifo.bFF_LoWatermark = (fifo.CPReadWriteDistance < fifo.CPLoWatermark);
|
|
|
|
|
|
|
|
bool bpInt = fifo.bFF_Breakpoint && fifo.bFF_BPInt;
|
|
|
|
bool ovfInt = fifo.bFF_HiWatermark && fifo.bFF_HiWatermarkInt;
|
|
|
|
bool undfInt = fifo.bFF_LoWatermark && fifo.bFF_LoWatermarkInt;
|
|
|
|
|
|
|
|
bool interrupt = (bpInt || ovfInt || undfInt) && m_CPCtrlReg.GPReadEnable;
|
|
|
|
|
|
|
|
if (interrupt != interruptSet && !interruptWaiting)
|
|
|
|
{
|
|
|
|
u64 userdata = interrupt ? 1 : 0;
|
|
|
|
if (IsOnThread())
|
|
|
|
{
|
|
|
|
if (!interrupt || bpInt || undfInt || ovfInt)
|
|
|
|
{
|
|
|
|
// Schedule the interrupt asynchronously
|
|
|
|
interruptWaiting = true;
|
|
|
|
CommandProcessor::UpdateInterruptsFromVideoBackend(userdata);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
CommandProcessor::UpdateInterrupts(userdata);
|
|
|
|
}
|
|
|
|
}
|
2014-08-24 20:27:32 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void SetCPStatusFromCPU()
|
|
|
|
{
|
|
|
|
// overflow & underflow check
|
|
|
|
fifo.bFF_HiWatermark = (fifo.CPReadWriteDistance > fifo.CPHiWatermark);
|
|
|
|
fifo.bFF_LoWatermark = (fifo.CPReadWriteDistance < fifo.CPLoWatermark);
|
2010-12-11 12:42:55 +00:00
|
|
|
|
|
|
|
bool bpInt = fifo.bFF_Breakpoint && fifo.bFF_BPInt;
|
|
|
|
bool ovfInt = fifo.bFF_HiWatermark && fifo.bFF_HiWatermarkInt;
|
|
|
|
bool undfInt = fifo.bFF_LoWatermark && fifo.bFF_LoWatermarkInt;
|
2013-03-14 06:08:26 +00:00
|
|
|
|
2010-12-13 04:44:33 +00:00
|
|
|
bool interrupt = (bpInt || ovfInt || undfInt) && m_CPCtrlReg.GPReadEnable;
|
2010-12-11 12:42:55 +00:00
|
|
|
|
2013-03-20 01:51:12 +00:00
|
|
|
if (interrupt != interruptSet && !interruptWaiting)
|
|
|
|
{
|
2014-11-14 06:07:11 +00:00
|
|
|
u64 userdata = interrupt ? 1 : 0;
|
2013-03-20 01:51:12 +00:00
|
|
|
if (IsOnThread())
|
|
|
|
{
|
|
|
|
if (!interrupt || bpInt || undfInt || ovfInt)
|
2012-03-05 05:40:10 +00:00
|
|
|
{
|
2014-11-14 00:46:02 +00:00
|
|
|
interruptSet = interrupt;
|
|
|
|
INFO_LOG(COMMANDPROCESSOR,"Interrupt set");
|
|
|
|
ProcessorInterface::SetInterrupt(INT_CAUSE_CP, interrupt);
|
2012-03-05 05:40:10 +00:00
|
|
|
}
|
2013-03-20 01:51:12 +00:00
|
|
|
}
|
|
|
|
else
|
2013-04-24 13:21:54 +00:00
|
|
|
{
|
2013-03-20 01:51:12 +00:00
|
|
|
CommandProcessor::UpdateInterrupts(userdata);
|
2013-04-24 13:21:54 +00:00
|
|
|
}
|
2013-03-20 01:51:12 +00:00
|
|
|
}
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
|
|
|
|
2010-12-13 07:56:54 +00:00
|
|
|
void ProcessFifoEvents()
|
|
|
|
{
|
2011-01-31 01:28:32 +00:00
|
|
|
if (IsOnThread() && (interruptWaiting || interruptFinishWaiting || interruptTokenWaiting))
|
|
|
|
CoreTiming::ProcessFifoWaitEvents();
|
2010-12-11 12:42:55 +00:00
|
|
|
}
|
|
|
|
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
void Shutdown()
|
|
|
|
{
|
2013-10-29 05:23:17 +00:00
|
|
|
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void SetCpStatusRegister()
|
|
|
|
{
|
|
|
|
// Here always there is one fifo attached to the GPU
|
|
|
|
m_CPStatusReg.Breakpoint = fifo.bFF_Breakpoint;
|
2014-12-28 00:03:13 +00:00
|
|
|
m_CPStatusReg.ReadIdle = !fifo.CPReadWriteDistance || (fifo.CPReadPointer == fifo.CPWritePointer);
|
2013-02-16 01:51:09 +00:00
|
|
|
m_CPStatusReg.CommandIdle = !fifo.CPReadWriteDistance || AtBreakpoint() || !fifo.bFF_GPReadEnable;
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
m_CPStatusReg.UnderflowLoWatermark = fifo.bFF_LoWatermark;
|
|
|
|
m_CPStatusReg.OverflowHiWatermark = fifo.bFF_HiWatermark;
|
2012-03-13 11:35:11 +00:00
|
|
|
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
INFO_LOG(COMMANDPROCESSOR,"\t Read from STATUS_REGISTER : %04x", m_CPStatusReg.Hex);
|
|
|
|
DEBUG_LOG(COMMANDPROCESSOR, "(r) status: iBP %s | fReadIdle %s | fCmdIdle %s | iOvF %s | iUndF %s"
|
2014-02-16 20:30:18 +00:00
|
|
|
, m_CPStatusReg.Breakpoint ? "ON" : "OFF"
|
|
|
|
, m_CPStatusReg.ReadIdle ? "ON" : "OFF"
|
|
|
|
, m_CPStatusReg.CommandIdle ? "ON" : "OFF"
|
|
|
|
, m_CPStatusReg.OverflowHiWatermark ? "ON" : "OFF"
|
|
|
|
, m_CPStatusReg.UnderflowLoWatermark ? "ON" : "OFF"
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
void SetCpControlRegister()
|
|
|
|
{
|
|
|
|
fifo.bFF_BPInt = m_CPCtrlReg.BPInt;
|
|
|
|
fifo.bFF_BPEnable = m_CPCtrlReg.BPEnable;
|
|
|
|
fifo.bFF_HiWatermarkInt = m_CPCtrlReg.FifoOverflowIntEnable;
|
|
|
|
fifo.bFF_LoWatermarkInt = m_CPCtrlReg.FifoUnderflowIntEnable;
|
|
|
|
fifo.bFF_GPLinkEnable = m_CPCtrlReg.GPLinkEnable;
|
|
|
|
|
2014-03-10 11:30:55 +00:00
|
|
|
if (fifo.bFF_GPReadEnable && !m_CPCtrlReg.GPReadEnable)
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
{
|
|
|
|
fifo.bFF_GPReadEnable = m_CPCtrlReg.GPReadEnable;
|
2015-03-05 20:14:46 +00:00
|
|
|
FlushGpu();
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
fifo.bFF_GPReadEnable = m_CPCtrlReg.GPReadEnable;
|
|
|
|
}
|
|
|
|
|
|
|
|
DEBUG_LOG(COMMANDPROCESSOR, "\t GPREAD %s | BP %s | Int %s | OvF %s | UndF %s | LINK %s"
|
2014-02-16 20:30:18 +00:00
|
|
|
, fifo.bFF_GPReadEnable ? "ON" : "OFF"
|
|
|
|
, fifo.bFF_BPEnable ? "ON" : "OFF"
|
|
|
|
, fifo.bFF_BPInt ? "ON" : "OFF"
|
|
|
|
, m_CPCtrlReg.FifoOverflowIntEnable ? "ON" : "OFF"
|
|
|
|
, m_CPCtrlReg.FifoUnderflowIntEnable ? "ON" : "OFF"
|
|
|
|
, m_CPCtrlReg.GPLinkEnable ? "ON" : "OFF"
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2014-06-01 08:56:09 +00:00
|
|
|
// NOTE: We intentionally don't emulate this function at the moment.
|
|
|
|
// We don't emulate proper GP timing anyway at the moment, so it would just slow down emulation.
|
Big Fifo Commit Part2: Now the fifo is more stable than my first commit, so is time...
- ReImplementing Single Core Mode like Dual Core Mode Style.
- Stage 1: My goal is, we have the Fifo, CommandProccessor code the more clear, maintenible and documented possible. When I quit dolphin I want any developer can continue with the work only reading the code.
* Big Refactoring: A lot of functions was changed the names, and modularized.
Now the FifoLoop and CatchUpGPU does not exist, was replaced by RunGpu() and RunGpuLoop().
The general idea is modeling the code like the real HW. The fifo is only a buffer where the Write Gather Pipe write the commands and from the Graphic Processor read these.
* Big Clean UP a lot of obsolete code and comments was deleted, like DcFakeWachDog, "Fifo very soon hack", etc.
In the stage 2, I will refactoring more code doing emphasis in the division of CommandProcessor, Fifo, Gpu Emulation. Beside I will comment all functions and variables in the code (Don't worry I will ask for English help for this part ;) )
Please test a lot SC mode and DC mode :)
Thank you so much for testing always and the patience. I don't like broke your favorite game but... you must believe me this part is very sensible, I only try to contribute for have a better and stable dolphin emulator.
git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@7185 8ced0084-cf51-0410-be5f-012b33b47a6e
2011-02-17 04:25:21 +00:00
|
|
|
void SetCpClearRegister()
|
|
|
|
{
|
|
|
|
}
|
2010-12-11 12:42:55 +00:00
|
|
|
|
2013-02-16 01:51:09 +00:00
|
|
|
void Update()
|
|
|
|
{
|
|
|
|
while (VITicks > m_cpClockOrigin && fifo.isGpuReadingData && IsOnThread())
|
|
|
|
Common::YieldCPU();
|
|
|
|
|
|
|
|
if (fifo.isGpuReadingData)
|
|
|
|
Common::AtomicAdd(VITicks, SystemTimers::GetTicksPerSecond() / 10000);
|
2015-03-05 16:12:24 +00:00
|
|
|
|
|
|
|
RunGpu();
|
2013-02-16 01:51:09 +00:00
|
|
|
}
|
2010-06-09 01:37:08 +00:00
|
|
|
} // end of namespace CommandProcessor
|