~80% speed improvement in decoding GX_TF_I8 textures. Yes, EIGHTY PERCENT. However, for MKWii movie playback I still can't break the fluffin' 48 FPS boundary on my machine! There's something else at play here because this decoder is ridonkulously fast.

~25% speed improvement in decoding GX_TF_RGB5A3 textures which aren't used very much. I thought it would help for movie playback but I misled myself. Video playback has nothing to do with this texture format.
Next I'll see if I can knock out some of these other texture decoders. Byte swizzling I'm sure can somehow be accomplished using _mm_unpacklo_epi8 trickery, so that'd be another big win I hope.

git-svn-id: https://dolphin-emu.googlecode.com/svn/trunk@6698 8ced0084-cf51-0410-be5f-012b33b47a6e
This commit is contained in:
james.jdunne 2010-12-31 07:23:17 +00:00
parent 41c40bcd50
commit 343e3f7c75
1 changed files with 287 additions and 31 deletions

View File

@ -972,38 +972,85 @@ PC_TexFormat TexDecoder_Decode_RGBA(u32 * dst, const u8 * src, int width, int he
break; break;
case GX_TF_I8: // speed critical case GX_TF_I8: // speed critical
{ {
#if _M_SSE >= 0x301
// JSD: It doesn't get any faster than this, folks.
for (int y = 0; y < height; y += 4) for (int y = 0; y < height; y += 4)
for (int x = 0; x < width; x += 8) for (int x = 0; x < width; x += 8)
#if _M_SSE >= 0x401
for (int iy = 0; iy < 4; ++iy, src += 8)
{ {
__m128i *quaddst = (__m128i *)(dst + (y + iy)*width + x); __m128i *quaddst;
const __m128i m0 = _mm_or_si128(
_mm_or_si128(
_mm_and_si128(_mm_set1_epi8(src[0]), _mm_set_epi32(0, 0, 0, (int)0xffffffffU)),
_mm_and_si128(_mm_set1_epi8(src[1]), _mm_set_epi32(0, 0, (int)0xffffffffU, 0))
),
_mm_or_si128(
_mm_and_si128(_mm_set1_epi8(src[2]), _mm_set_epi32(0, (int)0xffffffffU, 0, 0)),
_mm_and_si128(_mm_set1_epi8(src[3]), _mm_set_epi32((int)0xffffffffU, 0, 0, 0))
)
);
_mm_store_si128(quaddst, m0);
const __m128i m1 = _mm_or_si128( // Load 64 bits from `src` into an __m128i with upper 64 bits zeroed: (0000 0000 hgfe dcba)
_mm_or_si128( const __m128i r0 = _mm_loadl_epi64((const __m128i *)src);
_mm_and_si128(_mm_set1_epi8(src[4]), _mm_set_epi32(0, 0, 0, (int)0xffffffffU)), // Shuffle low 64-bits with itself to expand from (0000 0000 hgfe dcba) to (hhgg ffee ddcc bbaa)
_mm_and_si128(_mm_set1_epi8(src[5]), _mm_set_epi32(0, 0, (int)0xffffffffU, 0)) const __m128i r1 = _mm_unpacklo_epi8(r0, r0);
),
_mm_or_si128( // Shuffle low 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (dddd cccc bbbb aaaa)
_mm_and_si128(_mm_set1_epi8(src[6]), _mm_set_epi32(0, (int)0xffffffffU, 0, 0)), const __m128i rgba0 = _mm_unpacklo_epi8(r1, r1);
_mm_and_si128(_mm_set1_epi8(src[7]), _mm_set_epi32((int)0xffffffffU, 0, 0, 0)) // Shuffle hi 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (hhhh gggg ffff eeee)
) const __m128i rgba1 = _mm_unpackhi_epi8(r1, r1);
);
_mm_store_si128(quaddst+1, m1); // Store (dddd cccc bbbb aaaa) out:
quaddst = (__m128i *)(dst + (y + 0)*width + x);
_mm_store_si128(quaddst, rgba0);
// Store (hhhh gggg ffff eeee) out:
_mm_store_si128(quaddst+1, rgba1);
// Load 64 bits from `src` into an __m128i with upper 64 bits zeroed: (0000 0000 hgfe dcba)
src += 8;
const __m128i r2 = _mm_loadl_epi64((const __m128i *)src);
// Shuffle low 64-bits with itself to expand from (0000 0000 hgfe dcba) to (hhgg ffee ddcc bbaa)
const __m128i r3 = _mm_unpacklo_epi8(r2, r2);
// Shuffle low 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (dddd cccc bbbb aaaa)
const __m128i rgba2 = _mm_unpacklo_epi8(r3, r3);
// Shuffle hi 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (hhhh gggg ffff eeee)
const __m128i rgba3 = _mm_unpackhi_epi8(r3, r3);
// Store (dddd cccc bbbb aaaa) out:
quaddst = (__m128i *)(dst + (y + 1)*width + x);
_mm_store_si128(quaddst, rgba2);
// Store (hhhh gggg ffff eeee) out:
_mm_store_si128(quaddst+1, rgba3);
// Load 64 bits from `src` into an __m128i with upper 64 bits zeroed: (0000 0000 hgfe dcba)
src += 8;
const __m128i r4 = _mm_loadl_epi64((const __m128i *)src);
// Shuffle low 64-bits with itself to expand from (0000 0000 hgfe dcba) to (hhgg ffee ddcc bbaa)
const __m128i r5 = _mm_unpacklo_epi8(r4, r4);
// Shuffle low 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (dddd cccc bbbb aaaa)
const __m128i rgba4 = _mm_unpacklo_epi8(r5, r5);
// Shuffle hi 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (hhhh gggg ffff eeee)
const __m128i rgba5 = _mm_unpackhi_epi8(r5, r5);
// Store (dddd cccc bbbb aaaa) out:
quaddst = (__m128i *)(dst + (y + 2)*width + x);
_mm_store_si128(quaddst, rgba4);
// Store (hhhh gggg ffff eeee) out:
_mm_store_si128(quaddst+1, rgba5);
// Load 64 bits from `src` into an __m128i with upper 64 bits zeroed: (0000 0000 hgfe dcba)
src += 8;
const __m128i r6 = _mm_loadl_epi64((const __m128i *)src);
// Shuffle low 64-bits with itself to expand from (0000 0000 hgfe dcba) to (hhgg ffee ddcc bbaa)
const __m128i r7 = _mm_unpacklo_epi8(r6, r6);
// Shuffle low 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (dddd cccc bbbb aaaa)
const __m128i rgba6 = _mm_unpacklo_epi8(r7, r7);
// Shuffle hi 64-bits with itself to expand from (hhgg ffee ddcc bbaa) to (hhhh gggg ffff eeee)
const __m128i rgba7 = _mm_unpackhi_epi8(r7, r7);
// Store (dddd cccc bbbb aaaa) out:
quaddst = (__m128i *)(dst + (y + 3)*width + x);
_mm_store_si128(quaddst, rgba6);
// Store (hhhh gggg ffff eeee) out:
_mm_store_si128(quaddst+1, rgba7);
src += 8;
} }
#else #else
for (int y = 0; y < height; y += 4)
for (int x = 0; x < width; x += 8)
for (int iy = 0; iy < 4; ++iy, src += 8) for (int iy = 0; iy < 4; ++iy, src += 8)
{ {
u32 * newdst = dst + (y + iy)*width+x; u32 * newdst = dst + (y + iy)*width+x;
@ -1111,11 +1158,220 @@ PC_TexFormat TexDecoder_Decode_RGBA(u32 * dst, const u8 * src, int width, int he
} }
break; break;
case GX_TF_RGB5A3: case GX_TF_RGB5A3:
{ // JSD: speed critical for Mario Kart Wii intro movie (at least) {
#if _M_SSE >= 0x301
// These constants are used to apply the (x & mask) operation after x has been right-shifted
// out of its place.
const __m128i kMask_x1f = _mm_set_epi32(0x0000001fL, 0x0000001fL, 0x0000001fL, 0x0000001fL);
const __m128i kMask_x0f = _mm_set_epi32(0x0000000fL, 0x0000000fL, 0x0000000fL, 0x0000000fL);
const __m128i kMask_x07 = _mm_set_epi32(0x00000007L, 0x00000007L, 0x00000007L, 0x00000007L);
// This is the hard-coded 0xFF alpha constant that is ORed in place after the RGB are calculated
// for the RGB555 case when (s[x] & 0x8000) is true for all pixels.
const __m128i aVxff00 = _mm_set_epi32(0xFF000000L, 0xFF000000L, 0xFF000000L, 0xFF000000L);
for (int y = 0; y < height; y += 4)
for (int x = 0; x < width; x += 4)
for (int iy = 0; iy < 4; iy++, src += 8)
{
u32 *newdst = dst+(y+iy)*width+x;
const u16 *newsrc = (const u16*)src;
// TODO: weak point
const u16 val0 = Common::swap16(newsrc[0]);
const u16 val1 = Common::swap16(newsrc[1]);
const u16 val2 = Common::swap16(newsrc[2]);
const u16 val3 = Common::swap16(newsrc[3]);
// Need to check all 4 pixels' MSBs to ensure we can do data-parallelism:
if (((val0 & 0x8000) & (val1 & 0x8000) & (val2 & 0x8000) & (val3 & 0x8000)) == 0x8000)
{
// SSE2 case #1: all 4 pixels are in RGB555 and alpha = 0xFF.
const __m128i valV = _mm_set_epi16(0, val3, 0, val2, 0, val1, 0, val0);
// Swizzle bits: 00012345 -> 12345123
//r0 = (((val0>>10) & 0x1f) << 3) | (((val0>>10) & 0x1f) >> 2);
const __m128i tmprV = _mm_and_si128(_mm_srli_epi16(valV, 10), kMask_x1f);
const __m128i rV = _mm_or_si128( _mm_slli_epi16(tmprV, 3), _mm_srli_epi16(tmprV, 2) );
//newdst[0] = r0 | (_______) | (________) | (________);
__m128i final = rV;
//g0 = (((val0>>5 ) & 0x1f) << 3) | (((val0>>5 ) & 0x1f) >> 2);
const __m128i tmpgV = _mm_and_si128(_mm_srli_epi16(valV, 5), kMask_x1f);
const __m128i gV = _mm_or_si128( _mm_slli_epi16(tmpgV, 3), _mm_srli_epi16(tmpgV, 2) );
//newdst[0] = r0 | (g0 << 8) | (________) | (________);
final = _mm_or_si128(
final,
_mm_slli_epi32(gV, 8)
);
//b0 = (((val0 ) & 0x1f) << 3) | (((val0 ) & 0x1f) >> 2);
const __m128i tmpbV = _mm_and_si128(valV, kMask_x1f);
const __m128i bV = _mm_or_si128( _mm_slli_epi16(tmpbV, 3), _mm_srli_epi16(tmpbV, 2) );
//newdst[0] = r0 | (g0 << 8) | (b0 << 16) | (________);
final = _mm_or_si128(
final,
_mm_slli_epi32(bV, 16)
);
// Alphas are ORed in as a constant __m128i.
//a0 = 0xFF;
//newdst[0] = r0 | (g0 << 8) | (b0 << 16) | (a0 << 24);
final = _mm_or_si128(
final,
aVxff00
);
// write the final result:
_mm_store_si128( (__m128i*)newdst, final );
}
else if (((val0 & 0x8000) | (val1 & 0x8000) | (val2 & 0x8000) | (val3 & 0x8000)) == 0x0000)
{
// SSE2 case #2: all 4 pixels are in RGBA4443.
const __m128i valV = _mm_set_epi16(0, val3, 0, val2, 0, val1, 0, val0);
// Swizzle bits: 00001234 -> 12341234
//r0 = (((val0>>8 ) & 0xf) << 4) | ((val0>>8 ) & 0xf);
const __m128i tmprV = _mm_and_si128(_mm_srli_epi16(valV, 8), kMask_x0f);
const __m128i rV = _mm_or_si128( _mm_slli_epi16(tmprV, 4), tmprV );
//newdst[0] = r0 | (_______) | (________) | (________);
__m128i final = rV;
//g0 = (((val0>>4 ) & 0xf) << 4) | ((val0>>4 ) & 0xf);
const __m128i tmpgV = _mm_and_si128(_mm_srli_epi16(valV, 4), kMask_x0f);
const __m128i gV = _mm_or_si128( _mm_slli_epi16(tmpgV, 4), tmpgV );
//newdst[0] = r0 | (g0 << 8) | (________) | (________);
final = _mm_or_si128(
final,
_mm_slli_epi32(gV, 8)
);
//b0 = (((val0 ) & 0xf) << 4) | ((val0 ) & 0xf);
const __m128i tmpbV = _mm_and_si128(valV, kMask_x0f);
const __m128i bV = _mm_or_si128( _mm_slli_epi16(tmpbV, 4), tmpbV );
//newdst[0] = r0 | (g0 << 8) | (b0 << 16) | (________);
final = _mm_or_si128(
final,
_mm_slli_epi32(bV, 16)
);
//a0 = (((val0>>12) & 0x7) << 5) | (((val0>>12) & 0x7) << 2) | (((val0>>12) & 0x7) >> 1);
const __m128i tmpaV = _mm_and_si128(_mm_srli_epi16(valV, 12), kMask_x07);
const __m128i aV = _mm_or_si128(
_mm_slli_epi16(tmpaV, 5),
_mm_or_si128(
_mm_slli_epi16(tmpaV, 2),
_mm_srli_epi16(tmpaV, 1)
)
);
//newdst[0] = r0 | (g0 << 8) | (b0 << 16) | (a0 << 24);
final = _mm_or_si128(
final,
_mm_slli_epi32(aV, 24)
);
// write the final result:
_mm_store_si128( (__m128i*)newdst, final );
}
else
{
// Horrific fallback case, but hey at least it's inlined :D
// Maybe overkill? I see slight improvements on my machine as far as RDTSC
// counts and it's all done in registers (on x64). No temp memory moves!
int r0,g0,b0,a0;
int r1,g1,b1,a1;
int r2,g2,b2,a2;
int r3,g3,b3,a3;
// Normal operation, no parallelism to take advantage of:
if (val0 & 0x8000)
{
// Swizzle bits: 00012345 -> 12345123
r0 = (((val0>>10) & 0x1f) << 3) | (((val0>>10) & 0x1f) >> 2);
g0 = (((val0>>5 ) & 0x1f) << 3) | (((val0>>5 ) & 0x1f) >> 2);
b0 = (((val0 ) & 0x1f) << 3) | (((val0 ) & 0x1f) >> 2);
a0 = 0xFF;
}
else
{
a0 = (((val0>>12) & 0x7) << 5) | (((val0>>12) & 0x7) << 2) | (((val0>>12) & 0x7) >> 1);
// Swizzle bits: 00001234 -> 12341234
r0 = (((val0>>8 ) & 0xf) << 4) | ((val0>>8 ) & 0xf);
g0 = (((val0>>4 ) & 0xf) << 4) | ((val0>>4 ) & 0xf);
b0 = (((val0 ) & 0xf) << 4) | ((val0 ) & 0xf);
}
newdst[0] = r0 | (g0 << 8) | (b0 << 16) | (a0 << 24);
if (val1 & 0x8000)
{
// Swizzle bits: 00012345 -> 12345123
r1 = (((val1>>10) & 0x1f) << 3) | (((val1>>10) & 0x1f) >> 2);
g1 = (((val1>>5 ) & 0x1f) << 3) | (((val1>>5 ) & 0x1f) >> 2);
b1 = (((val1 ) & 0x1f) << 3) | (((val1 ) & 0x1f) >> 2);
a1 = 0xFF;
}
else
{
a1 = (((val1>>12) & 0x7) << 5) | (((val1>>12) & 0x7) << 2) | (((val1>>12) & 0x7) >> 1);
r1 = (((val1>>8 ) & 0xf) << 4) | ((val1>>8 ) & 0xf);
g1 = (((val1>>4 ) & 0xf) << 4) | ((val1>>4 ) & 0xf);
b1 = (((val1 ) & 0xf) << 4) | ((val1 ) & 0xf);
}
newdst[1] = r1 | (g1 << 8) | (b1 << 16) | (a1 << 24);
if (val2 & 0x8000)
{
// Swizzle bits: 00012345 -> 12345123
r2 = (((val2>>10) & 0x1f) << 3) | (((val2>>10) & 0x1f) >> 2);
g2 = (((val2>>5 ) & 0x1f) << 3) | (((val2>>5 ) & 0x1f) >> 2);
b2 = (((val2 ) & 0x1f) << 3) | (((val2 ) & 0x1f) >> 2);
a2 = 0xFF;
}
else
{
a2 = (((val2>>12) & 0x7) << 5) | (((val2>>12) & 0x7) << 2) | (((val2>>12) & 0x7) >> 1);
r2 = (((val2>>8 ) & 0xf) << 4) | ((val2>>8 ) & 0xf);
g2 = (((val2>>4 ) & 0xf) << 4) | ((val2>>4 ) & 0xf);
b2 = (((val2 ) & 0xf) << 4) | ((val2 ) & 0xf);
}
newdst[2] = r2 | (g2 << 8) | (b2 << 16) | (a2 << 24);
if (val3 & 0x8000)
{
// Swizzle bits: 00012345 -> 12345123
r3 = (((val3>>10) & 0x1f) << 3) | (((val3>>10) & 0x1f) >> 2);
g3 = (((val3>>5 ) & 0x1f) << 3) | (((val3>>5 ) & 0x1f) >> 2);
b3 = (((val3 ) & 0x1f) << 3) | (((val3 ) & 0x1f) >> 2);
a3 = 0xFF;
}
else
{
a3 = (((val3>>12) & 0x7) << 5) | (((val3>>12) & 0x7) << 2) | (((val3>>12) & 0x7) >> 1);
r3 = (((val3>>8 ) & 0xf) << 4) | ((val3>>8 ) & 0xf);
g3 = (((val3>>4 ) & 0xf) << 4) | ((val3>>4 ) & 0xf);
b3 = (((val3 ) & 0xf) << 4) | ((val3 ) & 0xf);
}
newdst[3] = r3 | (g3 << 8) | (b3 << 16) | (a3 << 24);
}
}
#else
for (int y = 0; y < height; y += 4) for (int y = 0; y < height; y += 4)
for (int x = 0; x < width; x += 4) for (int x = 0; x < width; x += 4)
for (int iy = 0; iy < 4; iy++, src += 8) for (int iy = 0; iy < 4; iy++, src += 8)
decodebytesRGB5A3rgba(dst+(y+iy)*width+x, (u16*)src); decodebytesRGB5A3rgba(dst+(y+iy)*width+x, (u16*)src);
#endif
} }
break; break;
case GX_TF_RGBA8: // speed critical case GX_TF_RGBA8: // speed critical