// Copyright (c) 2012- PPSSPP Project. // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, version 2.0 or later versions. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License 2.0 for more details. // A copy of the GPL 2.0 should have been included with the program. // If not, see http://www.gnu.org/licenses/ // Official git repository and contact information can be found at // https://github.com/hrydgard/ppsspp and http://www.ppsspp.org/. #include #include #include "math/math_util.h" #include "Core/MemMap.h" #include "Core/Config.h" #include "Core/Reporting.h" #include "Core/MIPS/MIPSAnalyst.h" #include "Core/MIPS/MIPSVFPUUtils.h" #include "Jit.h" #include "RegCache.h" // All functions should have CONDITIONAL_DISABLE, so we can narrow things down to a file quickly. // Currently known non working ones should have DISABLE. // #define CONDITIONAL_DISABLE { fpr.ReleaseSpillLocks(); Comp_Generic(op); return; } #define CONDITIONAL_DISABLE ; #define DISABLE { fpr.ReleaseSpillLocks(); Comp_Generic(op); return; } #define _RS ((op>>21) & 0x1F) #define _RT ((op>>16) & 0x1F) #define _RD ((op>>11) & 0x1F) #define _FS ((op>>11) & 0x1F) #define _FT ((op>>16) & 0x1F) #define _FD ((op>>6 ) & 0x1F) #define _POS ((op>>6 ) & 0x1F) #define _SIZE ((op>>11 ) & 0x1F) #ifndef M_LOG2E #define M_E 2.71828182845904523536f #define M_LOG2E 1.44269504088896340736f #define M_LOG10E 0.434294481903251827651f #define M_LN2 0.693147180559945309417f #define M_LN10 2.30258509299404568402f #undef M_PI #define M_PI 3.14159265358979323846f #define M_PI_2 1.57079632679489661923f #define M_PI_4 0.785398163397448309616f #define M_1_PI 0.318309886183790671538f #define M_2_PI 0.636619772367581343076f #define M_2_SQRTPI 1.12837916709551257390f #define M_SQRT2 1.41421356237309504880f #define M_SQRT1_2 0.707106781186547524401f #endif using namespace Gen; namespace MIPSComp { static const float one = 1.0f; static const float minus_one = -1.0f; static const float zero = 0.0f; const u32 GC_ALIGNED16( noSignMask[4] ) = {0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF, 0x7FFFFFFF}; const u32 GC_ALIGNED16( signBitLower[4] ) = {0x80000000, 0, 0, 0}; void Jit::Comp_VPFX(u32 op) { CONDITIONAL_DISABLE; int data = op & 0xFFFFF; int regnum = (op >> 24) & 3; switch (regnum) { case 0: // S js.prefixS = data; js.prefixSFlag = JitState::PREFIX_KNOWN_DIRTY; break; case 1: // T js.prefixT = data; js.prefixTFlag = JitState::PREFIX_KNOWN_DIRTY; break; case 2: // D js.prefixD = data; js.prefixDFlag = JitState::PREFIX_KNOWN_DIRTY; break; } } void Jit::ApplyPrefixST(u8 *vregs, u32 prefix, VectorSize sz) { if (prefix == 0xE4) return; int n = GetNumVectorElements(sz); u8 origV[4]; static const float constantArray[8] = {0.f, 1.f, 2.f, 0.5f, 3.f, 1.f/3.f, 0.25f, 1.f/6.f}; for (int i = 0; i < n; i++) origV[i] = vregs[i]; for (int i = 0; i < n; i++) { int regnum = (prefix >> (i*2)) & 3; int abs = (prefix >> (8+i)) & 1; int negate = (prefix >> (16+i)) & 1; int constants = (prefix >> (12+i)) & 1; // Unchanged, hurray. if (!constants && regnum == i && !abs && !negate) continue; // This puts the value into a temp reg, so we won't write the modified value back. vregs[i] = fpr.GetTempV(); fpr.MapRegV(vregs[i], MAP_NOINIT | MAP_DIRTY); if (!constants) { // Prefix may say "z, z, z, z" but if this is a pair, we force to x. // TODO: But some ops seem to use const 0 instead? if (regnum >= n) { ERROR_LOG_REPORT(CPU, "Invalid VFPU swizzle: %08x / %d", prefix, sz); regnum = 0; } MOVSS(fpr.VX(vregs[i]), fpr.V(origV[regnum])); if (abs) { ANDPS(fpr.VX(vregs[i]), M((void *)&noSignMask)); } } else { MOVSS(fpr.VX(vregs[i]), M((void *)&constantArray[regnum + (abs<<2)])); } if (negate) XORPS(fpr.VX(vregs[i]), M((void *)&signBitLower)); // TODO: This probably means it will swap out soon, inefficiently... fpr.ReleaseSpillLockV(vregs[i]); } } void Jit::GetVectorRegsPrefixD(u8 *regs, VectorSize sz, int vectorReg) { _assert_(js.prefixDFlag & JitState::PREFIX_KNOWN); GetVectorRegs(regs, sz, vectorReg); if (js.prefixD == 0) return; int n = GetNumVectorElements(sz); for (int i = 0; i < n; i++) { // Hopefully this is rare, we'll just write it into a reg we drop. if (js.VfpuWriteMask(i)) regs[i] = fpr.GetTempV(); } } void Jit::ApplyPrefixD(const u8 *vregs, VectorSize sz) { _assert_(js.prefixDFlag & JitState::PREFIX_KNOWN); if (!js.prefixD) return; int n = GetNumVectorElements(sz); for (int i = 0; i < n; i++) { if (js.VfpuWriteMask(i)) continue; int sat = (js.prefixD >> (i * 2)) & 3; if (sat == 1) { fpr.MapRegV(vregs[i], MAP_DIRTY); MAXSS(fpr.VX(vregs[i]), M((void *)&zero)); MINSS(fpr.VX(vregs[i]), M((void *)&one)); } else if (sat == 3) { fpr.MapRegV(vregs[i], MAP_DIRTY); MAXSS(fpr.VX(vregs[i]), M((void *)&minus_one)); MINSS(fpr.VX(vregs[i]), M((void *)&one)); } } } // Vector regs can overlap in all sorts of swizzled ways. // This does allow a single overlap in sregs[i]. bool IsOverlapSafeAllowS(int dreg, int di, int sn, u8 sregs[], int tn = 0, u8 tregs[] = NULL) { for (int i = 0; i < sn; ++i) { if (sregs[i] == dreg && i != di) return false; } for (int i = 0; i < tn; ++i) { if (tregs[i] == dreg) return false; } // Hurray, no overlap, we can write directly. return true; } bool IsOverlapSafe(int dreg, int di, int sn, u8 sregs[], int tn = 0, u8 tregs[] = NULL) { return IsOverlapSafeAllowS(dreg, di, sn, sregs, tn, tregs) && sregs[di] != dreg; } static u32 GC_ALIGNED16(ssLoadStoreTemp); void Jit::Comp_SV(u32 op) { CONDITIONAL_DISABLE; s32 imm = (signed short)(op&0xFFFC); int vt = ((op >> 16) & 0x1f) | ((op & 3) << 5); int rs = _RS; switch (op >> 26) { case 50: //lv.s // VI(vt) = Memory::Read_U32(addr); { gpr.BindToRegister(rs, true, false); fpr.MapRegV(vt, MAP_NOINIT); JitSafeMem safe(this, rs, imm); safe.SetFar(); OpArg src; if (safe.PrepareRead(src, 4)) { MOVSS(fpr.VX(vt), safe.NextFastAddress(0)); } if (safe.PrepareSlowRead((void *) &Memory::Read_U32)) { MOV(32, M((void *)&ssLoadStoreTemp), R(EAX)); MOVSS(fpr.VX(vt), M((void *)&ssLoadStoreTemp)); } safe.Finish(); gpr.UnlockAll(); fpr.ReleaseSpillLocks(); } break; case 58: //sv.s // Memory::Write_U32(VI(vt), addr); { gpr.BindToRegister(rs, true, true); // Even if we don't use real SIMD there's still 8 or 16 scalar float registers. fpr.MapRegV(vt, 0); JitSafeMem safe(this, rs, imm); safe.SetFar(); OpArg dest; if (safe.PrepareWrite(dest, 4)) { MOVSS(safe.NextFastAddress(0), fpr.VX(vt)); } if (safe.PrepareSlowWrite()) { MOVSS(M((void *)&ssLoadStoreTemp), fpr.VX(vt)); safe.DoSlowWrite((void *) &Memory::Write_U32, M((void *)&ssLoadStoreTemp), 0); } safe.Finish(); fpr.ReleaseSpillLocks(); gpr.UnlockAll(); } break; default: DISABLE; } } void Jit::Comp_SVQ(u32 op) { CONDITIONAL_DISABLE; int imm = (signed short)(op&0xFFFC); int vt = (((op >> 16) & 0x1f)) | ((op&1) << 5); int rs = _RS; switch (op >> 26) { case 54: //lv.q { gpr.BindToRegister(rs, true, true); u8 vregs[4]; GetVectorRegs(vregs, V_Quad, vt); fpr.MapRegsV(vregs, V_Quad, MAP_DIRTY | MAP_NOINIT); JitSafeMem safe(this, rs, imm); safe.SetFar(); OpArg src; if (safe.PrepareRead(src, 16)) { // Just copy 4 words the easiest way while not wasting registers. for (int i = 0; i < 4; i++) MOVSS(fpr.VX(vregs[i]), safe.NextFastAddress(i * 4)); } if (safe.PrepareSlowRead((void *) &Memory::Read_U32)) { for (int i = 0; i < 4; i++) { safe.NextSlowRead((void *) &Memory::Read_U32, i * 4); MOV(32, M((void *)&ssLoadStoreTemp), R(EAX)); MOVSS(fpr.VX(vregs[i]), M((void *)&ssLoadStoreTemp)); } } safe.Finish(); gpr.UnlockAll(); fpr.ReleaseSpillLocks(); } break; case 62: //sv.q { gpr.BindToRegister(rs, true, true); u8 vregs[4]; GetVectorRegs(vregs, V_Quad, vt); // Even if we don't use real SIMD there's still 8 or 16 scalar float registers. fpr.MapRegsV(vregs, V_Quad, 0); JitSafeMem safe(this, rs, imm); safe.SetFar(); OpArg dest; if (safe.PrepareWrite(dest, 16)) { for (int i = 0; i < 4; i++) MOVSS(safe.NextFastAddress(i * 4), fpr.VX(vregs[i])); } if (safe.PrepareSlowWrite()) { for (int i = 0; i < 4; i++) { MOVSS(M((void *)&ssLoadStoreTemp), fpr.VX(vregs[i])); safe.DoSlowWrite((void *) &Memory::Write_U32, M((void *)&ssLoadStoreTemp), i * 4); } } safe.Finish(); fpr.ReleaseSpillLocks(); gpr.UnlockAll(); } break; default: DISABLE; break; } } void Jit::Comp_VVectorInit(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; switch ((op >> 16) & 0xF) { case 6: // v=zeros; break; //vzero MOVSS(XMM0, M((void *) &zero)); break; case 7: // v=ones; break; //vone MOVSS(XMM0, M((void *) &one)); break; default: DISABLE; break; } VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); u8 dregs[4]; GetVectorRegsPrefixD(dregs, sz, _VD); fpr.MapRegsV(dregs, sz, MAP_NOINIT | MAP_DIRTY); for (int i = 0; i < n; ++i) MOVSS(fpr.VX(dregs[i]), R(XMM0)); ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } void Jit::Comp_VDot(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); // TODO: Force read one of them into regs? probably not. u8 sregs[4], tregs[4], dregs[1]; GetVectorRegsPrefixS(sregs, sz, _VS); GetVectorRegsPrefixT(tregs, sz, _VT); GetVectorRegsPrefixD(dregs, V_Single, _VD); X64Reg tempxreg = XMM0; if (IsOverlapSafe(dregs[0], 0, n, sregs, n, tregs)) { fpr.MapRegsV(dregs, V_Single, MAP_NOINIT); tempxreg = fpr.VX(dregs[0]); } // Need to start with +0.0f so it doesn't result in -0.0f. XORPS(tempxreg, R(tempxreg)); for (int i = 0; i < n; i++) { // sum += s[i]*t[i]; MOVSS(XMM1, fpr.V(sregs[i])); MULSS(XMM1, fpr.V(tregs[i])); ADDSS(tempxreg, R(XMM1)); } if (!fpr.V(dregs[0]).IsSimpleReg(tempxreg)) { fpr.MapRegsV(dregs, V_Single, MAP_NOINIT); MOVSS(fpr.V(dregs[0]), tempxreg); } ApplyPrefixD(dregs, V_Single); fpr.ReleaseSpillLocks(); } void Jit::Comp_VecDo3(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; void (XEmitter::*xmmop)(X64Reg, OpArg) = NULL; switch (op >> 26) { case 24: //VFPU0 switch ((op >> 23)&7) { case 0: // d[i] = s[i] + t[i]; break; //vadd xmmop = &XEmitter::ADDSS; break; case 1: // d[i] = s[i] - t[i]; break; //vsub xmmop = &XEmitter::SUBSS; break; case 7: // d[i] = s[i] / t[i]; break; //vdiv xmmop = &XEmitter::DIVSS; break; } break; case 25: //VFPU1 switch ((op >> 23) & 7) { case 0: // d[i] = s[i] * t[i]; break; //vmul xmmop = &XEmitter::MULSS; break; } break; case 27: //VFPU3 switch ((op >> 23) & 3) { case 2: // vmin xmmop = &XEmitter::MINSS; break; case 3: // vmax xmmop = &XEmitter::MAXSS; break; } break; default: _dbg_assert_msg_(CPU,0,"invalid VecDo3"); break; } if (xmmop == NULL) DISABLE; VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); u8 sregs[4], tregs[4], dregs[4]; GetVectorRegsPrefixS(sregs, sz, _VS); GetVectorRegsPrefixT(tregs, sz, _VT); GetVectorRegsPrefixD(dregs, sz, _VD); X64Reg tempxregs[4]; for (int i = 0; i < n; ++i) { if (!IsOverlapSafeAllowS(dregs[i], i, n, sregs, n, tregs)) { // On 32-bit we only have 6 xregs for mips regs, use XMM0/XMM1 if possible. if (i < 2) tempxregs[i] = (X64Reg) (XMM0 + i); else { int reg = fpr.GetTempV(); fpr.MapRegV(reg, MAP_NOINIT | MAP_DIRTY); fpr.SpillLockV(reg); tempxregs[i] = fpr.VX(reg); } } else { fpr.MapRegV(dregs[i], (dregs[i] == sregs[i] ? 0 : MAP_NOINIT) | MAP_DIRTY); fpr.SpillLockV(dregs[i]); tempxregs[i] = fpr.VX(dregs[i]); } } for (int i = 0; i < n; ++i) { if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); } for (int i = 0; i < n; ++i) (this->*xmmop)(tempxregs[i], fpr.V(tregs[i])); for (int i = 0; i < n; ++i) { if (!fpr.V(dregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(fpr.V(dregs[i]), tempxregs[i]); } ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } // There are no immediates for floating point, so we need to load these // from RAM. Might as well have a table ready. static const float mulTable[32] = { 1.0f/(1UL<<0),1.0f/(1UL<<1),1.0f/(1UL<<2),1.0f/(1UL<<3), 1.0f/(1UL<<4),1.0f/(1UL<<5),1.0f/(1UL<<6),1.0f/(1UL<<7), 1.0f/(1UL<<8),1.0f/(1UL<<9),1.0f/(1UL<<10),1.0f/(1UL<<11), 1.0f/(1UL<<12),1.0f/(1UL<<13),1.0f/(1UL<<14),1.0f/(1UL<<15), 1.0f/(1UL<<16),1.0f/(1UL<<17),1.0f/(1UL<<18),1.0f/(1UL<<19), 1.0f/(1UL<<20),1.0f/(1UL<<21),1.0f/(1UL<<22),1.0f/(1UL<<23), 1.0f/(1UL<<24),1.0f/(1UL<<25),1.0f/(1UL<<26),1.0f/(1UL<<27), 1.0f/(1UL<<28),1.0f/(1UL<<29),1.0f/(1UL<<30),1.0f/(1UL<<31), }; void Jit::Comp_Vi2f(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); int imm = (op >> 16) & 0x1f; const float *mult = &mulTable[imm]; u8 sregs[4], dregs[4]; GetVectorRegsPrefixS(sregs, sz, _VS); GetVectorRegsPrefixD(dregs, sz, _VD); MOVSS(XMM1, M((void *)mult)); for (int i = 0; i < n; i++) { if (fpr.V(sregs[i]).IsSimpleReg()) MOVD_xmm(R(EAX), fpr.VX(sregs[i])); else MOV(32, R(EAX), fpr.V(sregs[i])); CVTSI2SS(XMM0, R(EAX)); MULSS(XMM0, R(XMM1)); fpr.MapRegV(dregs[i], MAP_DIRTY); MOVSS(fpr.V(dregs[i]), XMM0); } ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } static const float cst_constants[32] = { 0, std::numeric_limits::max(), // all these are verified on real PSP sqrtf(2.0f), sqrtf(0.5f), 2.0f/sqrtf((float)M_PI), 2.0f/(float)M_PI, 1.0f/(float)M_PI, (float)M_PI/4, (float)M_PI/2, (float)M_PI, (float)M_E, (float)M_LOG2E, (float)M_LOG10E, (float)M_LN2, (float)M_LN10, 2*(float)M_PI, (float)M_PI/6, log10f(2.0f), logf(10.0f)/logf(2.0f), sqrtf(3.0f)/2.0f, }; void Jit::Comp_Vcst(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; int conNum = (op >> 16) & 0x1f; int vd = _VD; VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); u8 dregs[4]; GetVectorRegsPrefixD(dregs, sz, _VD); MOVSS(XMM0, M((void *)&cst_constants[conNum])); fpr.MapRegsV(dregs, sz, MAP_NOINIT | MAP_DIRTY); for (int i = 0; i < n; i++) { MOVSS(fpr.V(dregs[i]), XMM0); } ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } void Jit::Comp_VV2Op(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; // Pre-processing: Eliminate silly no-op VMOVs, common in Wipeout Pure if (((op >> 16) & 0x1f) == 0 && _VS == _VD && js.HasNoPrefix()) { return; } VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); u8 sregs[4], dregs[4]; GetVectorRegsPrefixS(sregs, sz, _VS); GetVectorRegsPrefixD(dregs, sz, _VD); X64Reg tempxregs[4]; for (int i = 0; i < n; ++i) { if (!IsOverlapSafeAllowS(dregs[i], i, n, sregs)) { int reg = fpr.GetTempV(); fpr.MapRegV(reg, MAP_NOINIT | MAP_DIRTY); fpr.SpillLockV(reg); tempxregs[i] = fpr.VX(reg); } else { fpr.MapRegV(dregs[i], (dregs[i] == sregs[i] ? 0 : MAP_NOINIT) | MAP_DIRTY); fpr.SpillLockV(dregs[i]); tempxregs[i] = fpr.VX(dregs[i]); } } // Warning: sregs[i] and tempxregs[i] may be the same reg. // Helps for vmov, hurts for vrcp, etc. for (int i = 0; i < n; ++i) { switch ((op >> 16) & 0x1f) { case 0: // d[i] = s[i]; break; //vmov // Probably for swizzle. if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); break; case 1: // d[i] = fabsf(s[i]); break; //vabs if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); ANDPS(tempxregs[i], M((void *)&noSignMask)); break; case 2: // d[i] = -s[i]; break; //vneg if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); XORPS(tempxregs[i], M((void *)&signBitLower)); break; case 4: // if (s[i] < 0) d[i] = 0; else {if(s[i] > 1.0f) d[i] = 1.0f; else d[i] = s[i];} break; // vsat0 if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); // TODO: Doesn't handle NaN correctly. MAXSS(tempxregs[i], M((void *)&zero)); MINSS(tempxregs[i], M((void *)&one)); break; case 5: // if (s[i] < -1.0f) d[i] = -1.0f; else {if(s[i] > 1.0f) d[i] = 1.0f; else d[i] = s[i];} break; // vsat1 if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); // TODO: Doesn't handle NaN correctly. MAXSS(tempxregs[i], M((void *)&minus_one)); MINSS(tempxregs[i], M((void *)&one)); break; case 16: // d[i] = 1.0f / s[i]; break; //vrcp MOVSS(XMM0, M((void *)&one)); DIVSS(XMM0, fpr.V(sregs[i])); MOVSS(tempxregs[i], R(XMM0)); break; case 17: // d[i] = 1.0f / sqrtf(s[i]); break; //vrsq SQRTSS(XMM0, fpr.V(sregs[i])); MOVSS(tempxregs[i], M((void *)&one)); DIVSS(tempxregs[i], R(XMM0)); break; case 18: // d[i] = sinf((float)M_PI_2 * s[i]); break; //vsin DISABLE; break; case 19: // d[i] = cosf((float)M_PI_2 * s[i]); break; //vcos DISABLE; break; case 20: // d[i] = powf(2.0f, s[i]); break; //vexp2 DISABLE; break; case 21: // d[i] = logf(s[i])/log(2.0f); break; //vlog2 DISABLE; break; case 22: // d[i] = sqrtf(s[i]); break; //vsqrt SQRTSS(tempxregs[i], fpr.V(sregs[i])); ANDPS(tempxregs[i], M((void *)&noSignMask)); break; case 23: // d[i] = asinf(s[i] * (float)M_2_PI); break; //vasin DISABLE; break; case 24: // d[i] = -1.0f / s[i]; break; // vnrcp MOVSS(XMM0, M((void *)&minus_one)); DIVSS(XMM0, fpr.V(sregs[i])); MOVSS(tempxregs[i], R(XMM0)); break; case 26: // d[i] = -sinf((float)M_PI_2 * s[i]); break; // vnsin DISABLE; break; case 28: // d[i] = 1.0f / expf(s[i] * (float)M_LOG2E); break; // vrexp2 DISABLE; break; } } for (int i = 0; i < n; ++i) { if (!fpr.V(dregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(fpr.V(dregs[i]), tempxregs[i]); } ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } void Jit::Comp_Mftv(u32 op) { CONDITIONAL_DISABLE; int imm = op & 0xFF; int rt = _RT; switch ((op >> 21) & 0x1f) { case 3: //mfv / mfvc // rt = 0, imm = 255 appears to be used as a CPU interlock by some games. if (rt != 0) { if (imm < 128) { //R(rt) = VI(imm); fpr.StoreFromRegisterV(imm); gpr.BindToRegister(rt, false, true); MOV(32, gpr.R(rt), fpr.V(imm)); } else if (imm < 128 + VFPU_CTRL_MAX) { //mtvc // In case we have a saved prefix. FlushPrefixV(); gpr.BindToRegister(rt, false, true); MOV(32, gpr.R(rt), M(¤tMIPS->vfpuCtrl[imm - 128])); } else { //ERROR - maybe need to make this value too an "interlock" value? _dbg_assert_msg_(CPU,0,"mfv - invalid register"); } } break; case 7: //mtv if (imm < 128) { fpr.StoreFromRegisterV(imm); gpr.BindToRegister(rt, true, false); MOV(32, fpr.V(imm), gpr.R(rt)); // VI(imm) = R(rt); } else if (imm < 128 + VFPU_CTRL_MAX) { //mtvc //currentMIPS->vfpuCtrl[imm - 128] = R(rt); gpr.BindToRegister(rt, true, false); MOV(32, M(¤tMIPS->vfpuCtrl[imm - 128]), gpr.R(rt)); // TODO: Optimization if rt is Imm? if (imm - 128 == VFPU_CTRL_SPREFIX) { js.prefixSFlag = JitState::PREFIX_UNKNOWN; } else if (imm - 128 == VFPU_CTRL_TPREFIX) { js.prefixTFlag = JitState::PREFIX_UNKNOWN; } else if (imm - 128 == VFPU_CTRL_DPREFIX) { js.prefixDFlag = JitState::PREFIX_UNKNOWN; } } else { //ERROR _dbg_assert_msg_(CPU,0,"mtv - invalid register"); } break; default: DISABLE; } } void Jit::Comp_Vmtvc(u32 op) { CONDITIONAL_DISABLE; int vs = _VS; int imm = op & 0xFF; if (imm >= 128 && imm < 128 + VFPU_CTRL_MAX) { fpr.MapRegV(vs, 0); MOVSS(M(¤tMIPS->vfpuCtrl[imm - 128]), fpr.VX(vs)); fpr.ReleaseSpillLocks(); if (imm - 128 == VFPU_CTRL_SPREFIX) { js.prefixSFlag = JitState::PREFIX_UNKNOWN; } else if (imm - 128 == VFPU_CTRL_TPREFIX) { js.prefixTFlag = JitState::PREFIX_UNKNOWN; } else if (imm - 128 == VFPU_CTRL_DPREFIX) { js.prefixDFlag = JitState::PREFIX_UNKNOWN; } } } void Jit::Comp_Vmmov(u32 op) { CONDITIONAL_DISABLE; // TODO: This probably ignores prefixes? if (js.MayHavePrefix()) DISABLE; MatrixSize sz = GetMtxSize(op); int n = GetMatrixSide(sz); u8 sregs[16], dregs[16]; GetMatrixRegs(sregs, sz, _VS); GetMatrixRegs(dregs, sz, _VD); // TODO: gas doesn't allow overlap, what does the PSP do? // Potentially detect overlap or the safe direction to move in, or just DISABLE? // This is very not optimal, blows the regcache everytime. u8 tempregs[16]; for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { u8 temp = (u8) fpr.GetTempV(); fpr.MapRegV(temp, MAP_NOINIT | MAP_DIRTY); MOVSS(fpr.VX(temp), fpr.V(sregs[a * 4 + b])); fpr.StoreFromRegisterV(temp); tempregs[a * 4 + b] = temp; } } for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { u8 temp = tempregs[a * 4 + b]; fpr.MapRegV(temp, 0); MOVSS(fpr.V(dregs[a * 4 + b]), fpr.VX(temp)); } } fpr.ReleaseSpillLocks(); } void Jit::Comp_VScl(u32 op) { CONDITIONAL_DISABLE; if (js.HasUnknownPrefix()) DISABLE; VectorSize sz = GetVecSize(op); int n = GetNumVectorElements(sz); u8 sregs[4], dregs[4], scale; GetVectorRegsPrefixS(sregs, sz, _VS); // TODO: Prefixes seem strange... GetVectorRegsPrefixT(&scale, V_Single, _VT); GetVectorRegsPrefixD(dregs, sz, _VD); // Move to XMM0 early, so we don't have to worry about overlap with scale. MOVSS(XMM0, fpr.V(scale)); X64Reg tempxregs[4]; for (int i = 0; i < n; ++i) { if (!IsOverlapSafeAllowS(dregs[i], i, n, sregs)) { int reg = fpr.GetTempV(); fpr.MapRegV(reg, MAP_NOINIT | MAP_DIRTY); fpr.SpillLockV(reg); tempxregs[i] = fpr.VX(reg); } else { fpr.MapRegV(dregs[i], (dregs[i] == sregs[i] ? 0 : MAP_NOINIT) | MAP_DIRTY); fpr.SpillLockV(dregs[i]); tempxregs[i] = fpr.VX(dregs[i]); } } for (int i = 0; i < n; ++i) { if (!fpr.V(sregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(tempxregs[i], fpr.V(sregs[i])); MULSS(tempxregs[i], R(XMM0)); } for (int i = 0; i < n; ++i) { if (!fpr.V(dregs[i]).IsSimpleReg(tempxregs[i])) MOVSS(fpr.V(dregs[i]), tempxregs[i]); } ApplyPrefixD(dregs, sz); fpr.ReleaseSpillLocks(); } void Jit::Comp_Vmmul(u32 op) { CONDITIONAL_DISABLE; // TODO: This probably ignores prefixes? if (js.MayHavePrefix()) DISABLE; MatrixSize sz = GetMtxSize(op); int n = GetMatrixSide(sz); u8 sregs[16], tregs[16], dregs[16]; GetMatrixRegs(sregs, sz, _VS); GetMatrixRegs(tregs, sz, _VT); GetMatrixRegs(dregs, sz, _VD); // TODO: test overlap, fix non-optimal. u8 tempregs[16]; for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { XORPS(XMM0, R(XMM0)); for (int c = 0; c < n; c++) { MOVSS(XMM1, fpr.V(sregs[b * 4 + c])); MULSS(XMM1, fpr.V(tregs[a * 4 + c])); ADDSS(XMM0, R(XMM1)); } u8 temp = (u8) fpr.GetTempV(); fpr.MapRegV(temp, MAP_NOINIT | MAP_DIRTY); MOVSS(fpr.VX(temp), R(XMM0)); fpr.StoreFromRegisterV(temp); tempregs[a * 4 + b] = temp; } } for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { u8 temp = tempregs[a * 4 + b]; fpr.MapRegV(temp, 0); MOVSS(fpr.V(dregs[a * 4 + b]), fpr.VX(temp)); } } fpr.ReleaseSpillLocks(); } void Jit::Comp_Vmscl(u32 op) { CONDITIONAL_DISABLE; // TODO: This probably ignores prefixes? if (js.MayHavePrefix()) DISABLE; MatrixSize sz = GetMtxSize(op); int n = GetMatrixSide(sz); u8 sregs[16], dregs[16], scale; GetMatrixRegs(sregs, sz, _VS); GetVectorRegs(&scale, V_Single, _VT); GetMatrixRegs(dregs, sz, _VD); // Move to XMM0 early, so we don't have to worry about overlap with scale. MOVSS(XMM0, fpr.V(scale)); // TODO: test overlap, optimize. u8 tempregs[16]; for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { u8 temp = (u8) fpr.GetTempV(); fpr.MapRegV(temp, MAP_NOINIT | MAP_DIRTY); MOVSS(fpr.VX(temp), fpr.V(sregs[a * 4 + b])); MULSS(fpr.VX(temp), R(XMM0)); fpr.StoreFromRegisterV(temp); tempregs[a * 4 + b] = temp; } } for (int a = 0; a < n; a++) { for (int b = 0; b < n; b++) { u8 temp = tempregs[a * 4 + b]; fpr.MapRegV(temp, 0); MOVSS(fpr.V(dregs[a * 4 + b]), fpr.VX(temp)); } } fpr.ReleaseSpillLocks(); } void Jit::Comp_Vtfm(u32 op) { CONDITIONAL_DISABLE; // TODO: This probably ignores prefixes? Or maybe uses D? if (js.MayHavePrefix()) DISABLE; VectorSize sz = GetVecSize(op); MatrixSize msz = GetMtxSize(op); int n = GetNumVectorElements(sz); int ins = (op >> 23) & 7; bool homogenous = false; if (n == ins) { n++; sz = (VectorSize)((int)(sz) + 1); msz = (MatrixSize)((int)(msz) + 1); homogenous = true; } // Otherwise, n should already be ins + 1. else if (n != ins + 1) DISABLE; u8 sregs[16], dregs[4], tregs[4]; GetMatrixRegs(sregs, msz, _VS); GetVectorRegs(tregs, sz, _VT); GetVectorRegs(dregs, sz, _VD); // TODO: test overlap, optimize. u8 tempregs[4]; for (int i = 0; i < n; i++) { XORPS(XMM0, R(XMM0)); for (int k = 0; k < n; k++) { MOVSS(XMM1, fpr.V(sregs[i * 4 + k])); if (!homogenous || k != n - 1) MULSS(XMM1, fpr.V(tregs[k])); ADDSS(XMM0, R(XMM1)); } u8 temp = (u8) fpr.GetTempV(); fpr.MapRegV(temp, MAP_NOINIT | MAP_DIRTY); MOVSS(fpr.VX(temp), R(XMM0)); fpr.StoreFromRegisterV(temp); tempregs[i] = temp; } for (int i = 0; i < n; i++) { u8 temp = tempregs[i]; fpr.MapRegV(temp, 0); MOVSS(fpr.V(dregs[i]), fpr.VX(temp)); } fpr.ReleaseSpillLocks(); } void Jit::Comp_VHdp(u32 op) { DISABLE; } void Jit::Comp_VCrs(u32 op) { DISABLE; } void Jit::Comp_VDet(u32 op) { DISABLE; } void Jit::Comp_Vi2x(u32 op) { DISABLE; } void Jit::Comp_Vx2i(u32 op) { DISABLE; } void Jit::Comp_Vf2i(u32 op) { DISABLE; } void Jit::Comp_Vhoriz(u32 op) { DISABLE; } }