mirror of https://github.com/PCSX2/pcsx2.git
1679 lines
47 KiB
C++
1679 lines
47 KiB
C++
/*
|
|
* Copyright (C) 2007-2009 Gabest
|
|
* http://www.gabest.org
|
|
*
|
|
* This Program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2, or (at your option)
|
|
* any later version.
|
|
*
|
|
* This Program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with GNU Make; see the file COPYING. If not, write to
|
|
* the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
|
|
* http://www.gnu.org/copyleft/gpl.html
|
|
*
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include "GS.h"
|
|
#include "GSTables.h"
|
|
#include "GSVector.h"
|
|
|
|
class GSBlock
|
|
{
|
|
static const GSVector4i m_r16mask;
|
|
static const GSVector4i m_r8mask;
|
|
static const GSVector4i m_r4mask;
|
|
|
|
static const GSVector4i m_xxxa;
|
|
static const GSVector4i m_xxbx;
|
|
static const GSVector4i m_xgxx;
|
|
static const GSVector4i m_rxxx;
|
|
|
|
static const GSVector4i m_uw8hmask0;
|
|
static const GSVector4i m_uw8hmask1;
|
|
static const GSVector4i m_uw8hmask2;
|
|
static const GSVector4i m_uw8hmask3;
|
|
|
|
public:
|
|
template<int i, bool aligned, uint32 mask> __forceinline static void WriteColumn32(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
const GSVector4i* s0 = (const GSVector4i*)&src[srcpitch * 0];
|
|
const GSVector4i* s1 = (const GSVector4i*)&src[srcpitch * 1];
|
|
|
|
GSVector4i v0 = GSVector4i::load<aligned>(&s0[0]);
|
|
GSVector4i v1 = GSVector4i::load<aligned>(&s0[1]);
|
|
GSVector4i v2 = GSVector4i::load<aligned>(&s1[0]);
|
|
GSVector4i v3 = GSVector4i::load<aligned>(&s1[1]);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
if(mask == 0xffffffff)
|
|
{
|
|
((GSVector4i*)dst)[i * 4 + 0] = v0;
|
|
((GSVector4i*)dst)[i * 4 + 1] = v1;
|
|
((GSVector4i*)dst)[i * 4 + 2] = v2;
|
|
((GSVector4i*)dst)[i * 4 + 3] = v3;
|
|
}
|
|
else
|
|
{
|
|
GSVector4i v4((int)mask);
|
|
|
|
#if _M_SSE >= 0x401
|
|
|
|
if(mask == 0xff000000 || mask == 0x00ffffff)
|
|
{
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend8(v0, v4);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend8(v1, v4);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend8(v2, v4);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend8(v3, v4);
|
|
}
|
|
else
|
|
{
|
|
|
|
#endif
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend(v0, v4);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend(v1, v4);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend(v2, v4);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend(v3, v4);
|
|
|
|
#if _M_SSE >= 0x401
|
|
|
|
}
|
|
|
|
#endif
|
|
}
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void WriteColumn16(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
const GSVector4i* s0 = (const GSVector4i*)&src[srcpitch * 0];
|
|
const GSVector4i* s1 = (const GSVector4i*)&src[srcpitch * 1];
|
|
|
|
GSVector4i v0 = GSVector4i::load<aligned>(&s0[0]);
|
|
GSVector4i v1 = GSVector4i::load<aligned>(&s0[1]);
|
|
GSVector4i v2 = GSVector4i::load<aligned>(&s1[0]);
|
|
GSVector4i v3 = GSVector4i::load<aligned>(&s1[1]);
|
|
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = v0;
|
|
((GSVector4i*)dst)[i * 4 + 1] = v2;
|
|
((GSVector4i*)dst)[i * 4 + 2] = v1;
|
|
((GSVector4i*)dst)[i * 4 + 3] = v3;
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void WriteColumn8(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
GSVector4i v0 = GSVector4i::load<aligned>(&src[srcpitch * 0]);
|
|
GSVector4i v1 = GSVector4i::load<aligned>(&src[srcpitch * 1]);
|
|
GSVector4i v2 = GSVector4i::load<aligned>(&src[srcpitch * 2]);
|
|
GSVector4i v3 = GSVector4i::load<aligned>(&src[srcpitch * 3]);
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
v2 = v2.yxwz();
|
|
v3 = v3.yxwz();
|
|
}
|
|
else
|
|
{
|
|
v0 = v0.yxwz();
|
|
v1 = v1.yxwz();
|
|
}
|
|
|
|
GSVector4i::sw8(v0, v2, v1, v3);
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = v0;
|
|
((GSVector4i*)dst)[i * 4 + 1] = v2;
|
|
((GSVector4i*)dst)[i * 4 + 2] = v1;
|
|
((GSVector4i*)dst)[i * 4 + 3] = v3;
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void WriteColumn4(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
// TODO: pshufb
|
|
|
|
GSVector4i v0 = GSVector4i::load<aligned>(&src[srcpitch * 0]);
|
|
GSVector4i v1 = GSVector4i::load<aligned>(&src[srcpitch * 1]);
|
|
GSVector4i v2 = GSVector4i::load<aligned>(&src[srcpitch * 2]);
|
|
GSVector4i v3 = GSVector4i::load<aligned>(&src[srcpitch * 3]);
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
v2 = v2.yxwzlh();
|
|
v3 = v3.yxwzlh();
|
|
}
|
|
else
|
|
{
|
|
v0 = v0.yxwzlh();
|
|
v1 = v1.yxwzlh();
|
|
}
|
|
|
|
GSVector4i::sw4(v0, v2, v1, v3);
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
GSVector4i::sw8(v0, v2, v1, v3);
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = v0;
|
|
((GSVector4i*)dst)[i * 4 + 1] = v1;
|
|
((GSVector4i*)dst)[i * 4 + 2] = v2;
|
|
((GSVector4i*)dst)[i * 4 + 3] = v3;
|
|
}
|
|
|
|
template<bool aligned, uint32 mask> static void WriteColumn32(int y, uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
switch((y >> 1) & 3)
|
|
{
|
|
case 0: WriteColumn32<0, aligned, mask>(dst, src, srcpitch); break;
|
|
case 1: WriteColumn32<1, aligned, mask>(dst, src, srcpitch); break;
|
|
case 2: WriteColumn32<2, aligned, mask>(dst, src, srcpitch); break;
|
|
case 3: WriteColumn32<3, aligned, mask>(dst, src, srcpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void WriteColumn16(int y, uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
switch((y >> 1) & 3)
|
|
{
|
|
case 0: WriteColumn16<0, aligned>(dst, src, srcpitch); break;
|
|
case 1: WriteColumn16<1, aligned>(dst, src, srcpitch); break;
|
|
case 2: WriteColumn16<2, aligned>(dst, src, srcpitch); break;
|
|
case 3: WriteColumn16<3, aligned>(dst, src, srcpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void WriteColumn8(int y, uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
switch((y >> 2) & 3)
|
|
{
|
|
case 0: WriteColumn8<0, aligned>(dst, src, srcpitch); break;
|
|
case 1: WriteColumn8<1, aligned>(dst, src, srcpitch); break;
|
|
case 2: WriteColumn8<2, aligned>(dst, src, srcpitch); break;
|
|
case 3: WriteColumn8<3, aligned>(dst, src, srcpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void WriteColumn4(int y, uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
switch((y >> 2) & 3)
|
|
{
|
|
case 0: WriteColumn4<0, aligned>(dst, src, srcpitch); break;
|
|
case 1: WriteColumn4<1, aligned>(dst, src, srcpitch); break;
|
|
case 2: WriteColumn4<2, aligned>(dst, src, srcpitch); break;
|
|
case 3: WriteColumn4<3, aligned>(dst, src, srcpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned, uint32 mask> static void WriteBlock32(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
WriteColumn32<0, aligned, mask>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn32<1, aligned, mask>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn32<2, aligned, mask>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn32<3, aligned, mask>(dst, src, srcpitch);
|
|
}
|
|
|
|
template<bool aligned> static void WriteBlock16(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
WriteColumn16<0, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn16<1, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn16<2, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 2;
|
|
WriteColumn16<3, aligned>(dst, src, srcpitch);
|
|
}
|
|
|
|
template<bool aligned> static void WriteBlock8(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
WriteColumn8<0, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn8<1, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn8<2, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn8<3, aligned>(dst, src, srcpitch);
|
|
}
|
|
|
|
template<bool aligned> static void WriteBlock4(uint8* RESTRICT dst, const uint8* RESTRICT src, int srcpitch)
|
|
{
|
|
WriteColumn4<0, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn4<1, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn4<2, aligned>(dst, src, srcpitch);
|
|
src += srcpitch * 4;
|
|
WriteColumn4<3, aligned>(dst, src, srcpitch);
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void ReadColumn32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0];
|
|
GSVector4i v1 = s[i * 4 + 1];
|
|
GSVector4i v2 = s[i * 4 + 2];
|
|
GSVector4i v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
GSVector4i* d0 = (GSVector4i*)&dst[dstpitch * 0];
|
|
GSVector4i* d1 = (GSVector4i*)&dst[dstpitch * 1];
|
|
|
|
GSVector4i::store<aligned>(&d0[0], v0);
|
|
GSVector4i::store<aligned>(&d0[1], v1);
|
|
GSVector4i::store<aligned>(&d1[0], v2);
|
|
GSVector4i::store<aligned>(&d1[1], v3);
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void ReadColumn16(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0].shuffle8(m_r16mask);
|
|
GSVector4i v1 = s[i * 4 + 1].shuffle8(m_r16mask);
|
|
GSVector4i v2 = s[i * 4 + 2].shuffle8(m_r16mask);
|
|
GSVector4i v3 = s[i * 4 + 3].shuffle8(m_r16mask);
|
|
|
|
GSVector4i::sw32(v0, v1, v2, v3);
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
GSVector4i* d0 = (GSVector4i*)&dst[dstpitch * 0];
|
|
GSVector4i* d1 = (GSVector4i*)&dst[dstpitch * 1];
|
|
|
|
GSVector4i::store<aligned>(&d0[0], v0);
|
|
GSVector4i::store<aligned>(&d0[1], v2);
|
|
GSVector4i::store<aligned>(&d1[0], v1);
|
|
GSVector4i::store<aligned>(&d1[1], v3);
|
|
|
|
#else
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0];
|
|
GSVector4i v1 = s[i * 4 + 1];
|
|
GSVector4i v2 = s[i * 4 + 2];
|
|
GSVector4i v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw32(v0, v1, v2, v3);
|
|
GSVector4i::sw16(v0, v2, v1, v3);
|
|
|
|
GSVector4i* d0 = (GSVector4i*)&dst[dstpitch * 0];
|
|
GSVector4i* d1 = (GSVector4i*)&dst[dstpitch * 1];
|
|
|
|
GSVector4i::store<aligned>(&d0[0], v0);
|
|
GSVector4i::store<aligned>(&d0[1], v1);
|
|
GSVector4i::store<aligned>(&d1[0], v2);
|
|
GSVector4i::store<aligned>(&d1[1], v3);
|
|
|
|
#endif
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void ReadColumn8(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
}
|
|
else
|
|
{
|
|
v2 = s[i * 4 + 0];
|
|
v3 = s[i * 4 + 1];
|
|
v0 = s[i * 4 + 2];
|
|
v1 = s[i * 4 + 3];
|
|
}
|
|
|
|
v0 = v0.shuffle8(m_r8mask);
|
|
v1 = v1.shuffle8(m_r8mask);
|
|
v2 = v2.shuffle8(m_r8mask);
|
|
v3 = v3.shuffle8(m_r8mask);
|
|
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw32(v0, v1, v3, v2);
|
|
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 0], v0);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 1], v3);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 2], v1);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 3], v2);
|
|
|
|
#else
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0];
|
|
GSVector4i v1 = s[i * 4 + 1];
|
|
GSVector4i v2 = s[i * 4 + 2];
|
|
GSVector4i v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw8(v0, v2, v1, v3);
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
v2 = v2.yxwz();
|
|
v3 = v3.yxwz();
|
|
}
|
|
else
|
|
{
|
|
v0 = v0.yxwz();
|
|
v1 = v1.yxwz();
|
|
}
|
|
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 0], v0);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 1], v1);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 2], v2);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 3], v3);
|
|
|
|
#endif
|
|
}
|
|
|
|
template<int i, bool aligned> __forceinline static void ReadColumn4(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0].xzyw();
|
|
GSVector4i v1 = s[i * 4 + 1].xzyw();
|
|
GSVector4i v2 = s[i * 4 + 2].xzyw();
|
|
GSVector4i v3 = s[i * 4 + 3].xzyw();
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
GSVector4i::sw4(v0, v2, v1, v3);
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
|
|
v0 = v0.shuffle8(m_r4mask);
|
|
v1 = v1.shuffle8(m_r4mask);
|
|
v2 = v2.shuffle8(m_r4mask);
|
|
v3 = v3.shuffle8(m_r4mask);
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
GSVector4i::sw16rh(v0, v1, v2, v3);
|
|
}
|
|
else
|
|
{
|
|
GSVector4i::sw16rl(v0, v1, v2, v3);
|
|
}
|
|
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 0], v0);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 1], v1);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 2], v2);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 3], v3);
|
|
|
|
#else
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = s[i * 4 + 0];
|
|
GSVector4i v1 = s[i * 4 + 1];
|
|
GSVector4i v2 = s[i * 4 + 2];
|
|
GSVector4i v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw32(v0, v1, v2, v3);
|
|
GSVector4i::sw32(v0, v1, v2, v3);
|
|
GSVector4i::sw4(v0, v2, v1, v3);
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
GSVector4i::sw16(v0, v2, v1, v3);
|
|
|
|
v0 = v0.xzyw();
|
|
v1 = v1.xzyw();
|
|
v2 = v2.xzyw();
|
|
v3 = v3.xzyw();
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
if((i & 1) == 0)
|
|
{
|
|
v2 = v2.yxwzlh();
|
|
v3 = v3.yxwzlh();
|
|
}
|
|
else
|
|
{
|
|
v0 = v0.yxwzlh();
|
|
v1 = v1.yxwzlh();
|
|
}
|
|
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 0], v0);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 1], v1);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 2], v2);
|
|
GSVector4i::store<aligned>(&dst[dstpitch * 3], v3);
|
|
|
|
#endif
|
|
}
|
|
|
|
template<bool aligned> static void ReadColumn32(int y, const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
switch((y >> 1) & 3)
|
|
{
|
|
case 0: ReadColumn32<0, aligned>(src, dst, dstpitch); break;
|
|
case 1: ReadColumn32<1, aligned>(src, dst, dstpitch); break;
|
|
case 2: ReadColumn32<2, aligned>(src, dst, dstpitch); break;
|
|
case 3: ReadColumn32<3, aligned>(src, dst, dstpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void ReadColumn16(int y, const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
switch((y >> 1) & 3)
|
|
{
|
|
case 0: ReadColumn16<0, aligned>(src, dst, dstpitch); break;
|
|
case 1: ReadColumn16<1, aligned>(src, dst, dstpitch); break;
|
|
case 2: ReadColumn16<2, aligned>(src, dst, dstpitch); break;
|
|
case 3: ReadColumn16<3, aligned>(src, dst, dstpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void ReadColumn8(int y, const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
switch((y >> 2) & 3)
|
|
{
|
|
case 0: ReadColumn8<0, aligned>(src, dst, dstpitch); break;
|
|
case 1: ReadColumn8<1, aligned>(src, dst, dstpitch); break;
|
|
case 2: ReadColumn8<2, aligned>(src, dst, dstpitch); break;
|
|
case 3: ReadColumn8<3, aligned>(src, dst, dstpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void ReadColumn4(int y, const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
switch((y >> 2) & 3)
|
|
{
|
|
case 0: ReadColumn4<0, aligned>(src, dst, dstpitch); break;
|
|
case 1: ReadColumn4<1, aligned>(src, dst, dstpitch); break;
|
|
case 2: ReadColumn4<2, aligned>(src, dst, dstpitch); break;
|
|
case 3: ReadColumn4<3, aligned>(src, dst, dstpitch); break;
|
|
default: __assume(0);
|
|
}
|
|
}
|
|
|
|
template<bool aligned> static void ReadBlock32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
ReadColumn32<0, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn32<1, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn32<2, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn32<3, aligned>(src, dst, dstpitch);
|
|
}
|
|
|
|
template<bool aligned> static void ReadBlock16(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
ReadColumn16<0, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn16<1, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn16<2, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 2;
|
|
ReadColumn16<3, aligned>(src, dst, dstpitch);
|
|
}
|
|
|
|
template<bool aligned> static void ReadBlock8(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
ReadColumn8<0, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn8<1, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn8<2, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn8<3, aligned>(src, dst, dstpitch);
|
|
}
|
|
|
|
template<bool aligned> static void ReadBlock4(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
ReadColumn4<0, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn4<1, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn4<2, aligned>(src, dst, dstpitch);
|
|
dst += dstpitch * 4;
|
|
ReadColumn4<3, aligned>(src, dst, dstpitch);
|
|
}
|
|
|
|
__forceinline static void ReadBlock4P(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
GSVector4i mask(0x0f0f0f0f);
|
|
|
|
for(int i = 0; i < 2; i++)
|
|
{
|
|
// col 0, 2
|
|
|
|
v0 = s[i * 8 + 0];
|
|
v1 = s[i * 8 + 1];
|
|
v2 = s[i * 8 + 2];
|
|
v3 = s[i * 8 + 3];
|
|
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw8(v0, v2, v1, v3);
|
|
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 0], (v0 & mask));
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 16], (v1 & mask));
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 0], (v2 & mask));
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 16], (v3 & mask));
|
|
|
|
dst += dstpitch * 2;
|
|
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 0], (v0.andnot(mask)).yxwz() >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 16], (v1.andnot(mask)).yxwz() >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 0], (v2.andnot(mask)).yxwz() >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 16], (v3.andnot(mask)).yxwz() >> 4);
|
|
|
|
dst += dstpitch * 2;
|
|
|
|
// col 1, 3
|
|
|
|
v0 = s[i * 8 + 4];
|
|
v1 = s[i * 8 + 5];
|
|
v2 = s[i * 8 + 6];
|
|
v3 = s[i * 8 + 7];
|
|
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw8(v0, v2, v1, v3);
|
|
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 0], (v0 & mask).yxwz());
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 16], (v1 & mask).yxwz());
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 0], (v2 & mask).yxwz());
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 16], (v3 & mask).yxwz());
|
|
|
|
dst += dstpitch * 2;
|
|
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 0], (v0.andnot(mask)) >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 0 + 16], (v1.andnot(mask)) >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 0], (v2.andnot(mask)) >> 4);
|
|
GSVector4i::store<true>(&dst[dstpitch * 1 + 16], (v3.andnot(mask)) >> 4);
|
|
|
|
dst += dstpitch * 2;
|
|
}
|
|
}
|
|
|
|
__forceinline static void ReadBlock8HP(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
v0 = ((v0 >> 24).ps32(v1 >> 24)).pu16((v2 >> 24).ps32(v3 >> 24));
|
|
|
|
GSVector4i::storel(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
|
|
GSVector4i::storeh(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
}
|
|
|
|
__forceinline static void ReadBlock4HLP(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
GSVector4i mask(0x0f0f0f0f);
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
v0 = ((v0 >> 24).ps32(v1 >> 24)).pu16((v2 >> 24).ps32(v3 >> 24)) & mask;
|
|
|
|
GSVector4i::storel(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
|
|
GSVector4i::storeh(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
}
|
|
|
|
__forceinline static void ReadBlock4HHP(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
v0 = ((v0 >> 28).ps32(v1 >> 28)).pu16((v2 >> 28).ps32(v3 >> 28));
|
|
|
|
GSVector4i::storel(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
|
|
GSVector4i::storeh(dst, v0);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
}
|
|
|
|
static void UnpackBlock24(const uint8* RESTRICT src, int srcpitch, uint32* RESTRICT dst)
|
|
{
|
|
GSVector4i mask = GSVector4i::x00ffffff();
|
|
|
|
for(int i = 0; i < 4; i++, src += srcpitch * 2)
|
|
{
|
|
GSVector4i v0 = GSVector4i::load<false>(src);
|
|
GSVector4i v1 = GSVector4i::load(src + 16, src + srcpitch);
|
|
GSVector4i v2 = GSVector4i::load<false>(src + srcpitch + 8);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = v0.upl32(v0.srl<3>()).upl64(v0.srl<6>().upl32(v0.srl<9>())) & mask;
|
|
|
|
v0 = v0.srl<12>(v1);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 1] = v0.upl32(v0.srl<3>()).upl64(v0.srl<6>().upl32(v0.srl<9>())) & mask;
|
|
|
|
v0 = v1.srl<8>(v2);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 2] = v0.upl32(v0.srl<3>()).upl64(v0.srl<6>().upl32(v0.srl<9>())) & mask;
|
|
|
|
v0 = v2.srl<4>();
|
|
|
|
((GSVector4i*)dst)[i * 4 + 3] = v0.upl32(v0.srl<3>()).upl64(v0.srl<6>().upl32(v0.srl<9>())) & mask;
|
|
}
|
|
}
|
|
|
|
static void UnpackBlock8H(const uint8* RESTRICT src, int srcpitch, uint32* RESTRICT dst)
|
|
{
|
|
GSVector4i zero = GSVector4i::zero();
|
|
|
|
for(int i = 0; i < 4; i++, src += srcpitch * 2)
|
|
{
|
|
GSVector4i v = GSVector4i::load(src, src + srcpitch);
|
|
|
|
GSVector4i v0 = zero.upl8(v);
|
|
GSVector4i v1 = zero.uph8(v);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = zero.upl16(v0);
|
|
((GSVector4i*)dst)[i * 4 + 1] = zero.uph16(v0);
|
|
((GSVector4i*)dst)[i * 4 + 2] = zero.upl16(v1);
|
|
((GSVector4i*)dst)[i * 4 + 3] = zero.uph16(v1);
|
|
}
|
|
}
|
|
|
|
static void UnpackBlock4HL(const uint8* RESTRICT src, int srcpitch, uint32* RESTRICT dst)
|
|
{
|
|
GSVector4i zero = GSVector4i::zero();
|
|
GSVector4i mask(0x0f0f0f0f);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = v & mask;
|
|
GSVector4i hi = (v >> 4) & mask;
|
|
|
|
GSVector4i v0 = lo.upl8(hi);
|
|
GSVector4i v1 = lo.uph8(hi);
|
|
|
|
GSVector4i v2 = zero.upl8(v0);
|
|
GSVector4i v3 = zero.uph8(v0);
|
|
GSVector4i v4 = zero.upl8(v1);
|
|
GSVector4i v5 = zero.uph8(v1);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = zero.upl16(v2);
|
|
((GSVector4i*)dst)[i * 8 + 1] = zero.uph16(v2);
|
|
((GSVector4i*)dst)[i * 8 + 2] = zero.upl16(v3);
|
|
((GSVector4i*)dst)[i * 8 + 3] = zero.uph16(v3);
|
|
((GSVector4i*)dst)[i * 8 + 4] = zero.upl16(v4);
|
|
((GSVector4i*)dst)[i * 8 + 5] = zero.uph16(v4);
|
|
((GSVector4i*)dst)[i * 8 + 6] = zero.upl16(v5);
|
|
((GSVector4i*)dst)[i * 8 + 7] = zero.uph16(v5);
|
|
}
|
|
}
|
|
|
|
static void UnpackBlock4HH(const uint8* RESTRICT src, int srcpitch, uint32* RESTRICT dst)
|
|
{
|
|
GSVector4i zero = GSVector4i::zero();
|
|
GSVector4i mask(0xf0f0f0f0);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = (v << 4) & mask;
|
|
GSVector4i hi = v & mask;
|
|
|
|
GSVector4i v0 = lo.upl8(hi);
|
|
GSVector4i v1 = lo.uph8(hi);
|
|
|
|
GSVector4i v2 = zero.upl8(v0);
|
|
GSVector4i v3 = zero.uph8(v0);
|
|
GSVector4i v4 = zero.upl8(v1);
|
|
GSVector4i v5 = zero.uph8(v1);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = zero.upl16(v2);
|
|
((GSVector4i*)dst)[i * 8 + 1] = zero.uph16(v2);
|
|
((GSVector4i*)dst)[i * 8 + 2] = zero.upl16(v3);
|
|
((GSVector4i*)dst)[i * 8 + 3] = zero.uph16(v3);
|
|
((GSVector4i*)dst)[i * 8 + 4] = zero.upl16(v4);
|
|
((GSVector4i*)dst)[i * 8 + 5] = zero.uph16(v4);
|
|
((GSVector4i*)dst)[i * 8 + 6] = zero.upl16(v5);
|
|
((GSVector4i*)dst)[i * 8 + 7] = zero.uph16(v5);
|
|
}
|
|
}
|
|
|
|
template<bool AEM> static void ExpandBlock24(const uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const GIFRegTEXA& TEXA)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i TA0(TEXA.TA0 << 24);
|
|
GSVector4i mask = GSVector4i::x00ffffff();
|
|
|
|
for(int i = 0; i < 4; i++, dst += dstpitch * 2)
|
|
{
|
|
GSVector4i v0 = s[i * 4 + 0] & mask;
|
|
GSVector4i v1 = s[i * 4 + 1] & mask;
|
|
GSVector4i v2 = s[i * 4 + 2] & mask;
|
|
GSVector4i v3 = s[i * 4 + 3] & mask;
|
|
|
|
GSVector4i* d0 = (GSVector4i*)&dst[dstpitch * 0];
|
|
GSVector4i* d1 = (GSVector4i*)&dst[dstpitch * 1];
|
|
|
|
if(AEM)
|
|
{
|
|
d0[0] = v0 | TA0.andnot(v0 == GSVector4i::zero()); // TA0 & (v0 != GSVector4i::zero())
|
|
d0[1] = v1 | TA0.andnot(v1 == GSVector4i::zero()); // TA0 & (v1 != GSVector4i::zero())
|
|
d1[0] = v2 | TA0.andnot(v2 == GSVector4i::zero()); // TA0 & (v2 != GSVector4i::zero())
|
|
d1[1] = v3 | TA0.andnot(v3 == GSVector4i::zero()); // TA0 & (v3 != GSVector4i::zero())
|
|
}
|
|
else
|
|
{
|
|
d0[0] = v0 | TA0;
|
|
d0[1] = v1 | TA0;
|
|
d1[0] = v2 | TA0;
|
|
d1[1] = v3 | TA0;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void ExpandBlock16(const uint16* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const GIFRegTEXA& TEXA) // do not inline, uses too many xmm regs
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i TA0(TEXA.TA0 << 24);
|
|
GSVector4i TA1(TEXA.TA1 << 24);
|
|
GSVector4i rm = m_rxxx;
|
|
GSVector4i gm = m_xgxx;
|
|
GSVector4i bm = m_xxbx;
|
|
GSVector4i l, h;
|
|
|
|
if(TEXA.AEM)
|
|
{
|
|
for(int i = 0; i < 8; i++, dst += dstpitch)
|
|
{
|
|
GSVector4i v0 = s[i * 2 + 0];
|
|
|
|
l = v0.upl16(v0);
|
|
h = v0.uph16(v0);
|
|
|
|
((GSVector4i*)dst)[0] = ((l & rm) << 3) | ((l & gm) << 6) | ((l & bm) << 9) | TA0.blend8(TA1, l.sra16(15)).andnot(l == GSVector4i::zero());
|
|
((GSVector4i*)dst)[1] = ((h & rm) << 3) | ((h & gm) << 6) | ((h & bm) << 9) | TA0.blend8(TA1, h.sra16(15)).andnot(h == GSVector4i::zero());
|
|
|
|
GSVector4i v1 = s[i * 2 + 1];
|
|
|
|
l = v1.upl16(v1);
|
|
h = v1.uph16(v1);
|
|
|
|
((GSVector4i*)dst)[2] = ((l & rm) << 3) | ((l & gm) << 6) | ((l & bm) << 9) | TA0.blend8(TA1, l.sra16(15)).andnot(l == GSVector4i::zero());
|
|
((GSVector4i*)dst)[3] = ((h & rm) << 3) | ((h & gm) << 6) | ((h & bm) << 9) | TA0.blend8(TA1, h.sra16(15)).andnot(h == GSVector4i::zero());
|
|
}
|
|
}
|
|
else
|
|
{
|
|
for(int i = 0; i < 8; i++, dst += dstpitch)
|
|
{
|
|
GSVector4i v0 = s[i * 2 + 0];
|
|
|
|
l = v0.upl16(v0);
|
|
h = v0.uph16(v0);
|
|
|
|
((GSVector4i*)dst)[0] = ((l & rm) << 3) | ((l & gm) << 6) | ((l & bm) << 9) | TA0.blend(TA1, l.sra16(15));
|
|
((GSVector4i*)dst)[1] = ((h & rm) << 3) | ((h & gm) << 6) | ((h & bm) << 9) | TA0.blend(TA1, h.sra16(15));
|
|
|
|
GSVector4i v1 = s[i * 2 + 1];
|
|
|
|
l = v1.upl16(v1);
|
|
h = v1.uph16(v1);
|
|
|
|
((GSVector4i*)dst)[2] = ((l & rm) << 3) | ((l & gm) << 6) | ((l & bm) << 9) | TA0.blend(TA1, l.sra16(15));
|
|
((GSVector4i*)dst)[3] = ((h & rm) << 3) | ((h & gm) << 6) | ((h & bm) << 9) | TA0.blend(TA1, h.sra16(15));
|
|
}
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock8_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 16; j++, dst += dstpitch)
|
|
{
|
|
((const GSVector4i*)src)[j].gather32_8(pal, (GSVector4i*)dst);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock8_16(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 16; j++, dst += dstpitch)
|
|
{
|
|
((const GSVector4i*)src)[j].gather16_8(pal, (GSVector4i*)dst);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const UINT64* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 16; j++, dst += dstpitch)
|
|
{
|
|
((const GSVector4i*)src)[j].gather64_8(pal, (GSVector4i*)dst);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4_16(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const UINT64* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 16; j++, dst += dstpitch)
|
|
{
|
|
((const GSVector4i*)src)[j].gather32_8(pal, (GSVector4i*)dst);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock8H_32(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
((GSVector4i*)dst)[0] = (s[j * 2 + 0] >> 24).gather32_32<>(pal);
|
|
((GSVector4i*)dst)[1] = (s[j * 2 + 1] >> 24).gather32_32<>(pal);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock8H_16(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = (s[j * 2 + 0] >> 24).gather32_32<>(pal);
|
|
GSVector4i v1 = (s[j * 2 + 1] >> 24).gather32_32<>(pal);
|
|
|
|
((GSVector4i*)dst)[0] = v0.pu32(v1);
|
|
|
|
#else
|
|
|
|
for(int i = 0; i < 8; i++)
|
|
{
|
|
((uint16*)dst)[i] = (uint16)pal[src[j * 8 + i] >> 24];
|
|
}
|
|
|
|
#endif
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4HL_32(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
((GSVector4i*)dst)[0] = ((s[j * 2 + 0] >> 24) & 0xf).gather32_32<>(pal);
|
|
((GSVector4i*)dst)[1] = ((s[j * 2 + 1] >> 24) & 0xf).gather32_32<>(pal);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4HL_16(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = ((s[j * 2 + 0] >> 24) & 0xf).gather32_32<>(pal);
|
|
GSVector4i v1 = ((s[j * 2 + 1] >> 24) & 0xf).gather32_32<>(pal);
|
|
|
|
((GSVector4i*)dst)[0] = v0.pu32(v1);
|
|
|
|
#else
|
|
|
|
for(int i = 0; i < 8; i++)
|
|
{
|
|
((uint16*)dst)[i] = (uint16)pal[(src[j * 8 + i] >> 24) & 0xf];
|
|
}
|
|
|
|
#endif
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4HH_32(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
((GSVector4i*)dst)[0] = (s[j * 2 + 0] >> 28).gather32_32<>(pal);
|
|
((GSVector4i*)dst)[1] = (s[j * 2 + 1] >> 28).gather32_32<>(pal);
|
|
}
|
|
}
|
|
|
|
__forceinline static void ExpandBlock4HH_16(uint32* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
for(int j = 0; j < 8; j++, dst += dstpitch)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0 = (s[j * 2 + 0] >> 28).gather32_32<>(pal);
|
|
GSVector4i v1 = (s[j * 2 + 1] >> 28).gather32_32<>(pal);
|
|
|
|
((GSVector4i*)dst)[0] = v0.pu32(v1);
|
|
|
|
#else
|
|
|
|
for(int i = 0; i < 8; i++)
|
|
{
|
|
((uint16*)dst)[i] = (uint16)pal[src[j * 8 + i] >> 28];
|
|
}
|
|
|
|
#endif
|
|
}
|
|
}
|
|
|
|
__forceinline static void UnpackAndWriteBlock24(const uint8* RESTRICT src, int srcpitch, uint8* RESTRICT dst)
|
|
{
|
|
GSVector4i mask(0x00ffffff);
|
|
|
|
for(int i = 0; i < 4; i++, src += srcpitch * 2)
|
|
{
|
|
GSVector4i v4 = GSVector4i::load<false>(src);
|
|
GSVector4i v5 = GSVector4i::load(src + 16, src + srcpitch);
|
|
GSVector4i v6 = GSVector4i::load<false>(src + srcpitch + 8);
|
|
|
|
GSVector4i v0 = v4.upl32(v4.srl<3>()).upl64(v4.srl<6>().upl32(v4.srl<9>()));
|
|
|
|
v4 = v4.srl<12>(v5);
|
|
|
|
GSVector4i v1 = v4.upl32(v4.srl<3>()).upl64(v4.srl<6>().upl32(v4.srl<9>()));
|
|
|
|
v4 = v5.srl<8>(v6);
|
|
|
|
GSVector4i v2 = v4.upl32(v4.srl<3>()).upl64(v4.srl<6>().upl32(v4.srl<9>()));
|
|
|
|
v4 = v6.srl<4>();
|
|
|
|
GSVector4i v3 = v4.upl32(v4.srl<3>()).upl64(v4.srl<6>().upl32(v4.srl<9>()));
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
#ifdef _M_AMD64
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend8(v0, mask);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend8(v1, mask);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend8(v2, mask);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend8(v3, mask);
|
|
|
|
#else
|
|
|
|
// here blend is faster than blend8 because vc8 has a little problem optimizing register usage for pblendvb (3rd op must be xmm0)
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend(v0, mask);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend(v1, mask);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend(v2, mask);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend(v3, mask);
|
|
|
|
#endif
|
|
}
|
|
}
|
|
|
|
__forceinline static void UnpackAndWriteBlock8H(const uint8* RESTRICT src, int srcpitch, uint8* RESTRICT dst)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
GSVector4i mask(0xff000000);
|
|
|
|
GSVector4i mask0 = m_uw8hmask0;
|
|
GSVector4i mask1 = m_uw8hmask1;
|
|
GSVector4i mask2 = m_uw8hmask2;
|
|
GSVector4i mask3 = m_uw8hmask3;
|
|
|
|
for(int i = 0; i < 4; i++, src += srcpitch * 2)
|
|
{
|
|
GSVector4i v4 = GSVector4i::load(src, src + srcpitch);
|
|
|
|
GSVector4i v0 = v4.shuffle8(mask0);
|
|
GSVector4i v1 = v4.shuffle8(mask1);
|
|
GSVector4i v2 = v4.shuffle8(mask2);
|
|
GSVector4i v3 = v4.shuffle8(mask3);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend8(v0, mask);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend8(v1, mask);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend8(v2, mask);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend8(v3, mask);
|
|
}
|
|
|
|
#else
|
|
|
|
GSVector4i mask(0xff000000);
|
|
|
|
for(int i = 0; i < 4; i++, src += srcpitch * 2)
|
|
{
|
|
GSVector4i v4 = GSVector4i::load(src, src + srcpitch);
|
|
|
|
GSVector4i v5 = v4.upl8(v4);
|
|
GSVector4i v6 = v4.uph8(v4);
|
|
|
|
GSVector4i v0 = v5.upl16(v5);
|
|
GSVector4i v1 = v5.uph16(v5);
|
|
GSVector4i v2 = v6.upl16(v6);
|
|
GSVector4i v3 = v6.uph16(v6);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 4 + 0] = ((GSVector4i*)dst)[i * 4 + 0].blend8(v0, mask);
|
|
((GSVector4i*)dst)[i * 4 + 1] = ((GSVector4i*)dst)[i * 4 + 1].blend8(v1, mask);
|
|
((GSVector4i*)dst)[i * 4 + 2] = ((GSVector4i*)dst)[i * 4 + 2].blend8(v2, mask);
|
|
((GSVector4i*)dst)[i * 4 + 3] = ((GSVector4i*)dst)[i * 4 + 3].blend8(v3, mask);
|
|
}
|
|
|
|
#endif
|
|
}
|
|
|
|
__forceinline static void UnpackAndWriteBlock4HL(const uint8* RESTRICT src, int srcpitch, uint8* RESTRICT dst)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
GSVector4i mask(0x0f0f0f0f);
|
|
GSVector4i mask0 = m_uw8hmask0;
|
|
GSVector4i mask1 = m_uw8hmask1;
|
|
GSVector4i mask2 = m_uw8hmask2;
|
|
GSVector4i mask3 = m_uw8hmask3;
|
|
GSVector4i mask4(0x0f000000);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = v & mask;
|
|
GSVector4i hi = (v >> 4) & mask;
|
|
|
|
{
|
|
GSVector4i v4 = lo.upl8(hi);
|
|
|
|
GSVector4i v0 = v4.shuffle8(mask0);
|
|
GSVector4i v1 = v4.shuffle8(mask1);
|
|
GSVector4i v2 = v4.shuffle8(mask2);
|
|
GSVector4i v3 = v4.shuffle8(mask3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = ((GSVector4i*)dst)[i * 8 + 0].blend(v0, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 1] = ((GSVector4i*)dst)[i * 8 + 1].blend(v1, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 2] = ((GSVector4i*)dst)[i * 8 + 2].blend(v2, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 3] = ((GSVector4i*)dst)[i * 8 + 3].blend(v3, mask4);
|
|
}
|
|
|
|
{
|
|
GSVector4i v4 = lo.uph8(hi);
|
|
|
|
GSVector4i v0 = v4.shuffle8(mask0);
|
|
GSVector4i v1 = v4.shuffle8(mask1);
|
|
GSVector4i v2 = v4.shuffle8(mask2);
|
|
GSVector4i v3 = v4.shuffle8(mask3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 4] = ((GSVector4i*)dst)[i * 8 + 4].blend(v0, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 5] = ((GSVector4i*)dst)[i * 8 + 5].blend(v1, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 6] = ((GSVector4i*)dst)[i * 8 + 6].blend(v2, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 7] = ((GSVector4i*)dst)[i * 8 + 7].blend(v3, mask4);
|
|
}
|
|
}
|
|
|
|
#else
|
|
/*
|
|
__declspec(align(16)) uint32 block[8 * 8];
|
|
|
|
UnpackBlock4HL(src, srcpitch, block);
|
|
|
|
WriteBlock32<true, 0x0f000000>(dst, (uint8*)block, sizeof(block) / 8);
|
|
*/
|
|
GSVector4i mask(0x0f0f0f0f);
|
|
GSVector4i mask2(0x0f000000);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = v & mask;
|
|
GSVector4i hi = (v >> 4) & mask;
|
|
|
|
{
|
|
GSVector4i v4 = lo.upl8(hi);
|
|
|
|
GSVector4i v5 = v4.upl8(v4);
|
|
GSVector4i v6 = v4.uph8(v4);
|
|
|
|
GSVector4i v0 = v5.upl16(v5);
|
|
GSVector4i v1 = v5.uph16(v5);
|
|
GSVector4i v2 = v6.upl16(v6);
|
|
GSVector4i v3 = v6.uph16(v6);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = ((GSVector4i*)dst)[i * 8 + 0].blend(v0, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 1] = ((GSVector4i*)dst)[i * 8 + 1].blend(v1, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 2] = ((GSVector4i*)dst)[i * 8 + 2].blend(v2, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 3] = ((GSVector4i*)dst)[i * 8 + 3].blend(v3, mask2);
|
|
}
|
|
|
|
{
|
|
GSVector4i v4 = lo.uph8(hi);
|
|
|
|
GSVector4i v5 = v4.upl8(v4);
|
|
GSVector4i v6 = v4.uph8(v4);
|
|
|
|
GSVector4i v0 = v5.upl16(v5);
|
|
GSVector4i v1 = v5.uph16(v5);
|
|
GSVector4i v2 = v6.upl16(v6);
|
|
GSVector4i v3 = v6.uph16(v6);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 4] = ((GSVector4i*)dst)[i * 8 + 4].blend(v0, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 5] = ((GSVector4i*)dst)[i * 8 + 5].blend(v1, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 6] = ((GSVector4i*)dst)[i * 8 + 6].blend(v2, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 7] = ((GSVector4i*)dst)[i * 8 + 7].blend(v3, mask2);
|
|
}
|
|
}
|
|
|
|
#endif
|
|
}
|
|
|
|
__forceinline static void UnpackAndWriteBlock4HH(const uint8* RESTRICT src, int srcpitch, uint8* RESTRICT dst)
|
|
{
|
|
#if _M_SSE >= 0x301
|
|
|
|
GSVector4i mask(0xf0f0f0f0);
|
|
GSVector4i mask0 = m_uw8hmask0;
|
|
GSVector4i mask1 = m_uw8hmask1;
|
|
GSVector4i mask2 = m_uw8hmask2;
|
|
GSVector4i mask3 = m_uw8hmask3;
|
|
GSVector4i mask4(0xf0000000);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = (v << 4) & mask;
|
|
GSVector4i hi = v & mask;
|
|
|
|
{
|
|
GSVector4i v4 = lo.upl8(hi);
|
|
|
|
GSVector4i v0 = v4.shuffle8(mask0);
|
|
GSVector4i v1 = v4.shuffle8(mask1);
|
|
GSVector4i v2 = v4.shuffle8(mask2);
|
|
GSVector4i v3 = v4.shuffle8(mask3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = ((GSVector4i*)dst)[i * 8 + 0].blend(v0, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 1] = ((GSVector4i*)dst)[i * 8 + 1].blend(v1, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 2] = ((GSVector4i*)dst)[i * 8 + 2].blend(v2, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 3] = ((GSVector4i*)dst)[i * 8 + 3].blend(v3, mask4);
|
|
}
|
|
|
|
{
|
|
GSVector4i v4 = lo.uph8(hi);
|
|
|
|
GSVector4i v0 = v4.shuffle8(mask0);
|
|
GSVector4i v1 = v4.shuffle8(mask1);
|
|
GSVector4i v2 = v4.shuffle8(mask2);
|
|
GSVector4i v3 = v4.shuffle8(mask3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 4] = ((GSVector4i*)dst)[i * 8 + 4].blend(v0, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 5] = ((GSVector4i*)dst)[i * 8 + 5].blend(v1, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 6] = ((GSVector4i*)dst)[i * 8 + 6].blend(v2, mask4);
|
|
((GSVector4i*)dst)[i * 8 + 7] = ((GSVector4i*)dst)[i * 8 + 7].blend(v3, mask4);
|
|
}
|
|
}
|
|
|
|
#else
|
|
/*
|
|
__declspec(align(16)) uint32 block[8 * 8];
|
|
|
|
UnpackBlock4HH(src, srcpitch, block);
|
|
|
|
WriteBlock32<true, 0xf0000000>(dst, (uint8*)block, sizeof(block) / 8);
|
|
*/
|
|
GSVector4i mask(0xf0f0f0f0);
|
|
GSVector4i mask2(0xf0000000);
|
|
|
|
for(int i = 0; i < 2; i++, src += srcpitch * 4)
|
|
{
|
|
GSVector4i v(
|
|
*(uint32*)&src[srcpitch * 0],
|
|
*(uint32*)&src[srcpitch * 1],
|
|
*(uint32*)&src[srcpitch * 2],
|
|
*(uint32*)&src[srcpitch * 3]);
|
|
|
|
GSVector4i lo = (v << 4) & mask;
|
|
GSVector4i hi = v & mask;
|
|
|
|
{
|
|
GSVector4i v4 = lo.upl8(hi);
|
|
|
|
GSVector4i v5 = v4.upl8(v4);
|
|
GSVector4i v6 = v4.uph8(v4);
|
|
|
|
GSVector4i v0 = v5.upl16(v5);
|
|
GSVector4i v1 = v5.uph16(v5);
|
|
GSVector4i v2 = v6.upl16(v6);
|
|
GSVector4i v3 = v6.uph16(v6);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 0] = ((GSVector4i*)dst)[i * 8 + 0].blend(v0, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 1] = ((GSVector4i*)dst)[i * 8 + 1].blend(v1, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 2] = ((GSVector4i*)dst)[i * 8 + 2].blend(v2, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 3] = ((GSVector4i*)dst)[i * 8 + 3].blend(v3, mask2);
|
|
}
|
|
|
|
{
|
|
GSVector4i v4 = lo.uph8(hi);
|
|
|
|
GSVector4i v5 = v4.upl8(v4);
|
|
GSVector4i v6 = v4.uph8(v4);
|
|
|
|
GSVector4i v0 = v5.upl16(v5);
|
|
GSVector4i v1 = v5.uph16(v5);
|
|
GSVector4i v2 = v6.upl16(v6);
|
|
GSVector4i v3 = v6.uph16(v6);
|
|
|
|
GSVector4i::sw64(v0, v2, v1, v3);
|
|
|
|
((GSVector4i*)dst)[i * 8 + 4] = ((GSVector4i*)dst)[i * 8 + 4].blend(v0, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 5] = ((GSVector4i*)dst)[i * 8 + 5].blend(v1, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 6] = ((GSVector4i*)dst)[i * 8 + 6].blend(v2, mask2);
|
|
((GSVector4i*)dst)[i * 8 + 7] = ((GSVector4i*)dst)[i * 8 + 7].blend(v3, mask2);
|
|
}
|
|
}
|
|
|
|
#endif
|
|
}
|
|
|
|
template<bool AEM> __forceinline static void ReadAndExpandBlock24(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const GIFRegTEXA& TEXA)
|
|
{
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i TA0(TEXA.TA0 << 24);
|
|
GSVector4i mask = GSVector4i::x00ffffff();
|
|
|
|
for(int i = 0; i < 4; i++, dst += dstpitch * 2)
|
|
{
|
|
GSVector4i v0 = s[i * 4 + 0];
|
|
GSVector4i v1 = s[i * 4 + 1];
|
|
GSVector4i v2 = s[i * 4 + 2];
|
|
GSVector4i v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
v0 &= mask;
|
|
v1 &= mask;
|
|
v2 &= mask;
|
|
v3 &= mask;
|
|
|
|
GSVector4i* d0 = (GSVector4i*)&dst[dstpitch * 0];
|
|
GSVector4i* d1 = (GSVector4i*)&dst[dstpitch * 1];
|
|
|
|
if(AEM)
|
|
{
|
|
d0[0] = v0 | TA0.andnot(v0 == GSVector4i::zero()); // TA0 & (v0 != GSVector4i::zero())
|
|
d0[1] = v1 | TA0.andnot(v1 == GSVector4i::zero()); // TA0 & (v1 != GSVector4i::zero())
|
|
d1[0] = v2 | TA0.andnot(v2 == GSVector4i::zero()); // TA0 & (v2 != GSVector4i::zero())
|
|
d1[1] = v3 | TA0.andnot(v3 == GSVector4i::zero()); // TA0 & (v3 != GSVector4i::zero())
|
|
}
|
|
else
|
|
{
|
|
d0[0] = v0 | TA0;
|
|
d0[1] = v1 | TA0;
|
|
d1[0] = v2 | TA0;
|
|
d1[1] = v3 | TA0;
|
|
}
|
|
}
|
|
}
|
|
|
|
__forceinline static void ReadAndExpandBlock8_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
GSVector4i mask = m_r8mask;
|
|
|
|
for(int i = 0; i < 2; i++)
|
|
{
|
|
v0 = s[i * 8 + 0].shuffle8(mask);
|
|
v1 = s[i * 8 + 1].shuffle8(mask);
|
|
v2 = s[i * 8 + 2].shuffle8(mask);
|
|
v3 = s[i * 8 + 3].shuffle8(mask);
|
|
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw32(v0, v1, v3, v2);
|
|
|
|
v0.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v3.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v1.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v2.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
|
|
v2 = s[i * 8 + 4].shuffle8(mask);
|
|
v3 = s[i * 8 + 5].shuffle8(mask);
|
|
v0 = s[i * 8 + 6].shuffle8(mask);
|
|
v1 = s[i * 8 + 7].shuffle8(mask);
|
|
|
|
GSVector4i::sw16(v0, v1, v2, v3);
|
|
GSVector4i::sw32(v0, v1, v3, v2);
|
|
|
|
v0.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v3.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v1.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v2.gather32_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
}
|
|
|
|
#else
|
|
|
|
__declspec(align(16)) uint8 block[16 * 16];
|
|
|
|
ReadBlock8<true>(src, (uint8*)block, sizeof(block) / 16);
|
|
|
|
ExpandBlock8_32(block, dst, dstpitch, pal);
|
|
|
|
#endif
|
|
}
|
|
|
|
// TODO: ReadAndExpandBlock8_16
|
|
|
|
__forceinline static void ReadAndExpandBlock4_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const UINT64* RESTRICT pal)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
GSVector4i mask = m_r4mask;
|
|
|
|
for(int i = 0; i < 2; i++)
|
|
{
|
|
v0 = s[i * 8 + 0].xzyw();
|
|
v1 = s[i * 8 + 1].xzyw();
|
|
v2 = s[i * 8 + 2].xzyw();
|
|
v3 = s[i * 8 + 3].xzyw();
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
GSVector4i::sw4(v0, v2, v1, v3);
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
|
|
v0 = v0.shuffle8(mask);
|
|
v1 = v1.shuffle8(mask);
|
|
v2 = v2.shuffle8(mask);
|
|
v3 = v3.shuffle8(mask);
|
|
|
|
GSVector4i::sw16rh(v0, v1, v2, v3);
|
|
|
|
v0.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v1.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v2.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v3.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
|
|
v0 = s[i * 8 + 4].xzyw();
|
|
v1 = s[i * 8 + 5].xzyw();
|
|
v2 = s[i * 8 + 6].xzyw();
|
|
v3 = s[i * 8 + 7].xzyw();
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
GSVector4i::sw4(v0, v2, v1, v3);
|
|
GSVector4i::sw8(v0, v1, v2, v3);
|
|
|
|
v0 = v0.shuffle8(mask);
|
|
v1 = v1.shuffle8(mask);
|
|
v2 = v2.shuffle8(mask);
|
|
v3 = v3.shuffle8(mask);
|
|
|
|
GSVector4i::sw16rl(v0, v1, v2, v3);
|
|
|
|
v0.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v1.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v2.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
v3.gather64_8<>(pal, (GSVector4i*)dst);
|
|
dst += dstpitch;
|
|
}
|
|
|
|
#else
|
|
|
|
__declspec(align(16)) uint8 block[(32 / 2) * 16];
|
|
|
|
ReadBlock4<true>(src, (uint8*)block, sizeof(block) / 16);
|
|
|
|
ExpandBlock4_32(block, dst, dstpitch, pal);
|
|
|
|
#endif
|
|
}
|
|
|
|
// TODO: ReadAndExpandBlock4_16
|
|
|
|
__forceinline static void ReadAndExpandBlock8H_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
(v0 >> 24).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
(v1 >> 24).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
|
|
(v2 >> 24).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
(v3 >> 24).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
|
|
#else
|
|
|
|
__declspec(align(16)) uint32 block[8 * 8];
|
|
|
|
ReadBlock32<true>(src, (uint8*)block, sizeof(block) / 8);
|
|
|
|
ExpandBlock8H_32(block, dst, dstpitch, pal);
|
|
|
|
#endif
|
|
}
|
|
|
|
// TODO: ReadAndExpandBlock8H_16
|
|
|
|
__forceinline static void ReadAndExpandBlock4HL_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
((v0 >> 24) & 0xf).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
((v1 >> 24) & 0xf).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
|
|
((v2 >> 24) & 0xf).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
((v3 >> 24) & 0xf).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
|
|
#else
|
|
|
|
__declspec(align(16)) uint32 block[8 * 8];
|
|
|
|
ReadBlock32<true>(src, (uint8*)block, sizeof(block) / 8);
|
|
|
|
ExpandBlock4HL_32(block, dst, dstpitch, pal);
|
|
|
|
#endif
|
|
}
|
|
|
|
// TODO: ReadAndExpandBlock4HL_16
|
|
|
|
__forceinline static void ReadAndExpandBlock4HH_32(const uint8* RESTRICT src, uint8* RESTRICT dst, int dstpitch, const uint32* RESTRICT pal)
|
|
{
|
|
#if _M_SSE >= 0x401
|
|
|
|
const GSVector4i* s = (const GSVector4i*)src;
|
|
|
|
GSVector4i v0, v1, v2, v3;
|
|
|
|
for(int i = 0; i < 4; i++)
|
|
{
|
|
v0 = s[i * 4 + 0];
|
|
v1 = s[i * 4 + 1];
|
|
v2 = s[i * 4 + 2];
|
|
v3 = s[i * 4 + 3];
|
|
|
|
GSVector4i::sw64(v0, v1, v2, v3);
|
|
|
|
(v0 >> 28).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
(v1 >> 28).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
|
|
(v2 >> 28).gather32_32<>(pal, (GSVector4i*)&dst[0]);
|
|
(v3 >> 28).gather32_32<>(pal, (GSVector4i*)&dst[16]);
|
|
|
|
dst += dstpitch;
|
|
}
|
|
|
|
#else
|
|
|
|
__declspec(align(16)) uint32 block[8 * 8];
|
|
|
|
ReadBlock32<true>(src, (uint8*)block, sizeof(block) / 8);
|
|
|
|
ExpandBlock4HH_32(block, dst, dstpitch, pal);
|
|
|
|
#endif
|
|
}
|
|
|
|
// TODO: ReadAndExpandBlock4HH_16
|
|
};
|