// Copyright 2008 Dolphin Emulator Project // Licensed under GPLv2+ // Refer to the license.txt file included. #include #include "Common/CommonFuncs.h" #include "Common/CPUDetect.h" #include "Common/Hash.h" #include "Common/Intrinsics.h" static u64 (*ptrHashFunction)(const u8 *src, u32 len, u32 samples) = &GetMurmurHash3; // uint32_t // WARNING - may read one more byte! // Implementation from Wikipedia. u32 HashFletcher(const u8* data_u8, size_t length) { const u16* data = (const u16*)data_u8; /* Pointer to the data to be summed */ size_t len = (length + 1) / 2; /* Length in 16-bit words */ u32 sum1 = 0xffff, sum2 = 0xffff; while (len) { size_t tlen = len > 360 ? 360 : len; len -= tlen; do { sum1 += *data++; sum2 += sum1; } while (--tlen); sum1 = (sum1 & 0xffff) + (sum1 >> 16); sum2 = (sum2 & 0xffff) + (sum2 >> 16); } // Second reduction step to reduce sums to 16 bits sum1 = (sum1 & 0xffff) + (sum1 >> 16); sum2 = (sum2 & 0xffff) + (sum2 >> 16); return(sum2 << 16 | sum1); } // Implementation from Wikipedia // Slightly slower than Fletcher above, but slightly more reliable. #define MOD_ADLER 65521 // data: Pointer to the data to be summed; len is in bytes u32 HashAdler32(const u8* data, size_t len) { u32 a = 1, b = 0; while (len) { size_t tlen = len > 5550 ? 5550 : len; len -= tlen; do { a += *data++; b += a; } while (--tlen); a = (a & 0xffff) + (a >> 16) * (65536 - MOD_ADLER); b = (b & 0xffff) + (b >> 16) * (65536 - MOD_ADLER); } // It can be shown that a <= 0x1013a here, so a single subtract will do. if (a >= MOD_ADLER) { a -= MOD_ADLER; } // It can be shown that b can reach 0xfff87 here. b = (b & 0xffff) + (b >> 16) * (65536 - MOD_ADLER); if (b >= MOD_ADLER) { b -= MOD_ADLER; } return((b << 16) | a); } // Stupid hash - but can't go back now :) // Don't use for new things. At least it's reasonably fast. u32 HashEctor(const u8* ptr, int length) { u32 crc = 0; for (int i = 0; i < length; i++) { crc ^= ptr[i]; crc = (crc << 3) | (crc >> 29); } return(crc); } #if _ARCH_64 //----------------------------------------------------------------------------- // Block read - if your platform needs to do endian-swapping or can only // handle aligned reads, do the conversion here inline u64 getblock(const u64 * p, int i) { return p[i]; } //---------- // Block mix - combine the key bits with the hash bits and scramble everything inline void bmix64(u64 & h1, u64 & h2, u64 & k1, u64 & k2, u64 & c1, u64 & c2) { k1 *= c1; k1 = _rotl64(k1, 23); k1 *= c2; h1 ^= k1; h1 += h2; h2 = _rotl64(h2, 41); k2 *= c2; k2 = _rotl64(k2, 23); k2 *= c1; h2 ^= k2; h2 += h1; h1 = h1 * 3 + 0x52dce729; h2 = h2 * 3 + 0x38495ab5; c1 = c1 * 5 + 0x7b7d159c; c2 = c2 * 5 + 0x6bce6396; } //---------- // Finalization mix - avalanches all bits to within 0.05% bias inline u64 fmix64(u64 k) { k ^= k >> 33; k *= 0xff51afd7ed558ccd; k ^= k >> 33; k *= 0xc4ceb9fe1a85ec53; k ^= k >> 33; return k; } u64 GetMurmurHash3(const u8 *src, u32 len, u32 samples) { const u8 * data = (const u8*)src; const int nblocks = len / 16; u32 Step = (len / 8); if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; u64 h1 = 0x9368e53c2f6af274; u64 h2 = 0x586dcd208f7cd3fd; u64 c1 = 0x87c37b91114253d5; u64 c2 = 0x4cf5ad432745937f; //---------- // body const u64 * blocks = (const u64 *)(data); for (int i = 0; i < nblocks; i+=Step) { u64 k1 = getblock(blocks, i*2+0); u64 k2 = getblock(blocks, i*2+1); bmix64(h1,h2,k1,k2,c1,c2); } //---------- // tail const u8 * tail = (const u8*)(data + nblocks*16); u64 k1 = 0; u64 k2 = 0; switch (len & 15) { case 15: k2 ^= u64(tail[14]) << 48; case 14: k2 ^= u64(tail[13]) << 40; case 13: k2 ^= u64(tail[12]) << 32; case 12: k2 ^= u64(tail[11]) << 24; case 11: k2 ^= u64(tail[10]) << 16; case 10: k2 ^= u64(tail[ 9]) << 8; case 9: k2 ^= u64(tail[ 8]) << 0; case 8: k1 ^= u64(tail[ 7]) << 56; case 7: k1 ^= u64(tail[ 6]) << 48; case 6: k1 ^= u64(tail[ 5]) << 40; case 5: k1 ^= u64(tail[ 4]) << 32; case 4: k1 ^= u64(tail[ 3]) << 24; case 3: k1 ^= u64(tail[ 2]) << 16; case 2: k1 ^= u64(tail[ 1]) << 8; case 1: k1 ^= u64(tail[ 0]) << 0; bmix64(h1,h2,k1,k2,c1,c2); }; //---------- // finalization h2 ^= len; h1 += h2; h2 += h1; h1 = fmix64(h1); h2 = fmix64(h2); h1 += h2; return h1; } // CRC32 hash using the SSE4.2 instruction u64 GetCRC32(const u8 *src, u32 len, u32 samples) { #if _M_SSE >= 0x402 || defined(_M_ARM_64) u64 h[4] = { len, 0, 0, 0 }; u32 Step = (len / 8); const u64 *data = (const u64 *)src; const u64 *end = data + Step; if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; #endif #if _M_SSE >= 0x402 while (data < end - Step * 3) { h[0] = _mm_crc32_u64(h[0], data[Step * 0]); h[1] = _mm_crc32_u64(h[1], data[Step * 1]); h[2] = _mm_crc32_u64(h[2], data[Step * 2]); h[3] = _mm_crc32_u64(h[3], data[Step * 3]); data += Step * 4; } if (data < end - Step * 0) h[0] = _mm_crc32_u64(h[0], data[Step * 0]); if (data < end - Step * 1) h[1] = _mm_crc32_u64(h[1], data[Step * 1]); if (data < end - Step * 2) h[2] = _mm_crc32_u64(h[2], data[Step * 2]); if (len & 7) { u64 temp = 0; memcpy(&temp, end, len & 7); h[0] = _mm_crc32_u64(h[0], temp); } #elif defined(_M_ARM_64) // We should be able to use intrinsics for this // Too bad the intrinsics for this instruction was added in GCC 4.9.1 // The Android NDK (as of r10e) only has GCC 4.9 // Once the Android NDK has a newer GCC version, update these to use intrinsics while (data < end - Step * 3) { asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[0]) : [two] "r" (h[0]), [three] "r" (data[Step * 0])); asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[1]) : [two] "r" (h[1]), [three] "r" (data[Step * 1])); asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[2]) : [two] "r" (h[2]), [three] "r" (data[Step * 2])); asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[3]) : [two] "r" (h[3]), [three] "r" (data[Step * 3])); data += Step * 4; } if (data < end - Step * 0) asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[0]) : [two] "r" (h[0]), [three] "r" (data[Step * 0])); if (data < end - Step * 1) asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[1]) : [two] "r" (h[1]), [three] "r" (data[Step * 1])); if (data < end - Step * 2) asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[2]) : [two] "r" (h[2]), [three] "r" (data[Step * 2])); if (len & 7) { u64 temp = 0; memcpy(&temp, end, len & 7); asm ("crc32x %w[res], %w[two], %x[three]" : [res] "=r" (h[0]) : [two] "r" (h[0]), [three] "r" (temp)); } #endif #if _M_SSE >= 0x402 || defined(_M_ARM_64) // FIXME: is there a better way to combine these partial hashes? return h[0] + (h[1] << 10) + (h[2] << 21) + (h[3] << 32); #else return 0; #endif } /* * NOTE: This hash function is used for custom texture loading/dumping, so * it should not be changed, which would require all custom textures to be * recalculated for their new hash values. If the hashing function is * changed, make sure this one is still used when the legacy parameter is * true. */ u64 GetHashHiresTexture(const u8 *src, u32 len, u32 samples) { const u64 m = 0xc6a4a7935bd1e995; u64 h = len * m; const int r = 47; u32 Step = (len / 8); const u64 *data = (const u64 *)src; const u64 *end = data + Step; if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; while (data < end) { u64 k = data[0]; data+=Step; k *= m; k ^= k >> r; k *= m; h ^= k; h *= m; } const u8 * data2 = (const u8*)end; switch (len & 7) { case 7: h ^= u64(data2[6]) << 48; case 6: h ^= u64(data2[5]) << 40; case 5: h ^= u64(data2[4]) << 32; case 4: h ^= u64(data2[3]) << 24; case 3: h ^= u64(data2[2]) << 16; case 2: h ^= u64(data2[1]) << 8; case 1: h ^= u64(data2[0]); h *= m; }; h ^= h >> r; h *= m; h ^= h >> r; return h; } #else // CRC32 hash using the SSE4.2 instruction u64 GetCRC32(const u8 *src, u32 len, u32 samples) { #if _M_SSE >= 0x402 u32 h = len; u32 Step = (len/4); const u32 *data = (const u32 *)src; const u32 *end = data + Step; if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; while (data < end) { h = _mm_crc32_u32(h, data[0]); data += Step; } const u8 *data2 = (const u8*)end; return (u64)_mm_crc32_u32(h, u32(data2[0])); #else return 0; #endif } //----------------------------------------------------------------------------- // Block read - if your platform needs to do endian-swapping or can only // handle aligned reads, do the conversion here inline u32 getblock(const u32 * p, int i) { return p[i]; } //---------- // Finalization mix - force all bits of a hash block to avalanche // avalanches all bits to within 0.25% bias inline u32 fmix32(u32 h) { h ^= h >> 16; h *= 0x85ebca6b; h ^= h >> 13; h *= 0xc2b2ae35; h ^= h >> 16; return h; } inline void bmix32(u32 & h1, u32 & h2, u32 & k1, u32 & k2, u32 & c1, u32 & c2) { k1 *= c1; k1 = _rotl(k1, 11); k1 *= c2; h1 ^= k1; h1 += h2; h2 = _rotl(h2, 17); k2 *= c2; k2 = _rotl(k2, 11); k2 *= c1; h2 ^= k2; h2 += h1; h1 = h1*3+0x52dce729; h2 = h2*3+0x38495ab5; c1 = c1*5+0x7b7d159c; c2 = c2*5+0x6bce6396; } //---------- u64 GetMurmurHash3(const u8* src, u32 len, u32 samples) { const u8 * data = (const u8*)src; u32 out[2]; const int nblocks = len / 8; u32 Step = (len / 4); if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; u32 h1 = 0x8de1c3ac; u32 h2 = 0xbab98226; u32 c1 = 0x95543787; u32 c2 = 0x2ad7eb25; //---------- // body const u32 * blocks = (const u32 *)(data + nblocks*8); for (int i = -nblocks; i < 0; i+=Step) { u32 k1 = getblock(blocks,i*2+0); u32 k2 = getblock(blocks,i*2+1); bmix32(h1,h2,k1,k2,c1,c2); } //---------- // tail const u8 * tail = (const u8*)(data + nblocks*8); u32 k1 = 0; u32 k2 = 0; switch (len & 7) { case 7: k2 ^= tail[6] << 16; case 6: k2 ^= tail[5] << 8; case 5: k2 ^= tail[4] << 0; case 4: k1 ^= tail[3] << 24; case 3: k1 ^= tail[2] << 16; case 2: k1 ^= tail[1] << 8; case 1: k1 ^= tail[0] << 0; bmix32(h1,h2,k1,k2,c1,c2); }; //---------- // finalization h2 ^= len; h1 += h2; h2 += h1; h1 = fmix32(h1); h2 = fmix32(h2); h1 += h2; h2 += h1; out[0] = h1; out[1] = h2; return *((u64 *)&out); } /* * FIXME: The old 32-bit version of this hash made different hashes than the * 64-bit version. Until someone can make a new version of the 32-bit one that * makes identical hashes, this is just a c/p of the 64-bit one. */ u64 GetHashHiresTexture(const u8 *src, u32 len, u32 samples) { const u64 m = 0xc6a4a7935bd1e995ULL; u64 h = len * m; const int r = 47; u32 Step = (len / 8); const u64 *data = (const u64 *)src; const u64 *end = data + Step; if (samples == 0) samples = std::max(Step, 1u); Step = Step / samples; if (Step < 1) Step = 1; while (data < end) { u64 k = data[0]; data+=Step; k *= m; k ^= k >> r; k *= m; h ^= k; h *= m; } const u8 * data2 = (const u8*)end; switch (len & 7) { case 7: h ^= u64(data2[6]) << 48; case 6: h ^= u64(data2[5]) << 40; case 5: h ^= u64(data2[4]) << 32; case 4: h ^= u64(data2[3]) << 24; case 3: h ^= u64(data2[2]) << 16; case 2: h ^= u64(data2[1]) << 8; case 1: h ^= u64(data2[0]); h *= m; }; h ^= h >> r; h *= m; h ^= h >> r; return h; } #endif u64 GetHash64(const u8 *src, u32 len, u32 samples) { return ptrHashFunction(src, len, samples); } // sets the hash function used for the texture cache void SetHash64Function() { #if _M_SSE >= 0x402 if (cpu_info.bSSE4_2) // sse crc32 version { ptrHashFunction = &GetCRC32; } else #elif defined(_M_ARM_64) if (cpu_info.bCRC32) { ptrHashFunction = &GetCRC32; } else #endif { ptrHashFunction = &GetMurmurHash3; } }