wyhash.h 6.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145
  1. /* Copyright 2020 王一 Wang Yi <godspeed_china@yeah.net>
  2. This is free and unencumbered software released into the public domain. http://unlicense.org/
  3. See github.com/wangyi-fudan/wyhash/ LICENSE
  4. */
  5. #ifndef wyhash_final_version
  6. #define wyhash_final_version
  7. //defines that change behavior
  8. #ifndef WYHASH_CONDOM
  9. #define WYHASH_CONDOM 1 //0: read 8 bytes before and after boundaries, dangerous but fastest. 1: normal valid behavior 2: extra protection against entropy loss (probability=2^-63), aka. "blind multiplication"
  10. #endif
  11. #define WYHASH_32BIT_MUM 0 //faster on 32 bit system
  12. //includes
  13. #include <stdint.h>
  14. #include <string.h>
  15. #if defined(_MSC_VER) && defined(_M_X64)
  16. #include <intrin.h>
  17. #pragma intrinsic(_umul128)
  18. #endif
  19. #if defined(__GNUC__) || defined(__INTEL_COMPILER) || defined(__clang__)
  20. #define _likely_(x) __builtin_expect(x,1)
  21. #define _unlikely_(x) __builtin_expect(x,0)
  22. #else
  23. #define _likely_(x) (x)
  24. #define _unlikely_(x) (x)
  25. #endif
  26. //mum function
  27. static inline uint64_t _wyrot(uint64_t x) { return (x>>32)|(x<<32); }
  28. static inline void _wymum(uint64_t *A, uint64_t *B){
  29. #if(WYHASH_32BIT_MUM)
  30. uint64_t hh=(*A>>32)*(*B>>32), hl=(*A>>32)*(unsigned)*B, lh=(unsigned)*A*(*B>>32), ll=(uint64_t)(unsigned)*A*(unsigned)*B;
  31. #if(WYHASH_CONDOM>1)
  32. *A^=_wyrot(hl)^hh; *B^=_wyrot(lh)^ll;
  33. #else
  34. *A=_wyrot(hl)^hh; *B=_wyrot(lh)^ll;
  35. #endif
  36. #elif defined(__SIZEOF_INT128__)
  37. __uint128_t r=*A; r*=*B;
  38. #if(WYHASH_CONDOM>1)
  39. *A^=(uint64_t)r; *B^=(uint64_t)(r>>64);
  40. #else
  41. *A=(uint64_t)r; *B=(uint64_t)(r>>64);
  42. #endif
  43. #elif defined(_MSC_VER) && defined(_M_X64)
  44. #if(WYHASH_CONDOM>1)
  45. uint64_t a, b;
  46. a=_umul128(*A,*B,&b);
  47. *A^=a; *B^=b;
  48. #else
  49. *A=_umul128(*A,*B,B);
  50. #endif
  51. #else
  52. uint64_t ha=*A>>32, hb=*B>>32, la=(uint32_t)*A, lb=(uint32_t)*B, hi, lo;
  53. uint64_t rh=ha*hb, rm0=ha*lb, rm1=hb*la, rl=la*lb, t=rl+(rm0<<32), c=t<rl;
  54. lo=t+(rm1<<32); c+=lo<t; hi=rh+(rm0>>32)+(rm1>>32)+c;
  55. #if(WYHASH_CONDOM>1)
  56. *A^=lo; *B^=hi;
  57. #else
  58. *A=lo; *B=hi;
  59. #endif
  60. #endif
  61. }
  62. static inline uint64_t _wymix(uint64_t A, uint64_t B){ _wymum(&A,&B); return A^B; }
  63. //read functions
  64. #ifndef WYHASH_LITTLE_ENDIAN
  65. #if defined(_WIN32) || defined(__LITTLE_ENDIAN__) || (defined(__BYTE_ORDER__) && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__)
  66. #define WYHASH_LITTLE_ENDIAN 1
  67. #elif defined(__BIG_ENDIAN__) || (defined(__BYTE_ORDER__) && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__)
  68. #define WYHASH_LITTLE_ENDIAN 0
  69. #endif
  70. #endif
  71. #if (WYHASH_LITTLE_ENDIAN)
  72. static inline uint64_t _wyr8(const uint8_t *p) { uint64_t v; memcpy(&v, p, 8); return v;}
  73. static inline uint64_t _wyr4(const uint8_t *p) { unsigned v; memcpy(&v, p, 4); return v;}
  74. #elif defined(__GNUC__) || defined(__INTEL_COMPILER) || defined(__clang__)
  75. static inline uint64_t _wyr8(const uint8_t *p) { uint64_t v; memcpy(&v, p, 8); return __builtin_bswap64(v);}
  76. static inline uint64_t _wyr4(const uint8_t *p) { unsigned v; memcpy(&v, p, 4); return __builtin_bswap32(v);}
  77. #elif defined(_MSC_VER)
  78. static inline uint64_t _wyr8(const uint8_t *p) { uint64_t v; memcpy(&v, p, 8); return _byteswap_uint64(v);}
  79. static inline uint64_t _wyr4(const uint8_t *p) { unsigned v; memcpy(&v, p, 4); return _byteswap_ulong(v);}
  80. #endif
  81. static inline uint64_t _wyr3(const uint8_t *p, unsigned k) { return (((uint64_t)p[0])<<16)|(((uint64_t)p[k>>1])<<8)|p[k-1];}
  82. //wyhash function
  83. static inline uint64_t _wyfinish16(const uint8_t *p, uint64_t len, uint64_t seed, const uint64_t *secret, uint64_t i){
  84. #if(WYHASH_CONDOM>0)
  85. uint64_t a, b;
  86. if(_likely_(i<=8)){
  87. if(_likely_(i>=4)){ a=_wyr4(p); b=_wyr4(p+i-4); }
  88. else if (_likely_(i)){ a=_wyr3(p,i); b=0; }
  89. else a=b=0;
  90. }
  91. else{ a=_wyr8(p); b=_wyr8(p+i-8); }
  92. return _wymix(secret[1]^len,_wymix(a^secret[1], b^seed));
  93. #else
  94. #define oneshot_shift ((i<8)*((8-i)<<3))
  95. return _wymix(secret[1]^len,_wymix((_wyr8(p)<<oneshot_shift)^secret[1],(_wyr8(p+i-8)>>oneshot_shift)^seed));
  96. #endif
  97. }
  98. static inline uint64_t _wyfinish(const uint8_t *p, uint64_t len, uint64_t seed, const uint64_t *secret, uint64_t i){
  99. if(_likely_(i<=16)) return _wyfinish16(p,len,seed,secret,i);
  100. return _wyfinish(p+16,len,_wymix(_wyr8(p)^secret[1],_wyr8(p+8)^seed),secret,i-16);
  101. }
  102. static inline uint64_t wyhash(const void *key, uint64_t len, uint64_t seed, const uint64_t *secret){
  103. const uint8_t *p=(const uint8_t *)key;
  104. uint64_t i=len; seed^=*secret;
  105. if(_unlikely_(i>64)){
  106. uint64_t see1=seed;
  107. do{
  108. seed=_wymix(_wyr8(p)^secret[1],_wyr8(p+8)^seed)^_wymix(_wyr8(p+16)^secret[2],_wyr8(p+24)^seed);
  109. see1=_wymix(_wyr8(p+32)^secret[3],_wyr8(p+40)^see1)^_wymix(_wyr8(p+48)^secret[4],_wyr8(p+56)^see1);
  110. p+=64; i-=64;
  111. }while(i>64);
  112. seed^=see1;
  113. }
  114. return _wyfinish(p,len,seed,secret,i);
  115. }
  116. //utility functions
  117. static const uint64_t _wyp[5] = {0xa0761d6478bd642full, 0xe7037ed1a0b428dbull, 0x8ebc6af09c88c6e3ull, 0x589965cc75374cc3ull, 0x1d8e4e27c47d124full};
  118. static inline uint64_t wyhash64(uint64_t A, uint64_t B){ A^=_wyp[0]; B^=_wyp[1]; _wymum(&A,&B); return _wymix(A^_wyp[0],B^_wyp[1]);}
  119. static inline uint64_t wyrand(uint64_t *seed){ *seed+=_wyp[0]; return _wymix(*seed,*seed^_wyp[1]);}
  120. static inline double wy2u01(uint64_t r){ const double _wynorm=1.0/(1ull<<52); return (r>>12)*_wynorm;}
  121. static inline double wy2gau(uint64_t r){ const double _wynorm=1.0/(1ull<<20); return ((r&0x1fffff)+((r>>21)&0x1fffff)+((r>>42)&0x1fffff))*_wynorm-3.0;}
  122. static inline uint64_t wy2u0k(uint64_t r, uint64_t k){ _wymum(&r,&k); return k; }
  123. static inline void make_secret(uint64_t seed, uint64_t *secret){
  124. uint8_t c[] = {15, 23, 27, 29, 30, 39, 43, 45, 46, 51, 53, 54, 57, 58, 60, 71, 75, 77, 78, 83, 85, 86, 89, 90, 92, 99, 101, 102, 105, 106, 108, 113, 114, 116, 120, 135, 139, 141, 142, 147, 149, 150, 153, 154, 156, 163, 165, 166, 169, 170, 172, 177, 178, 180, 184, 195, 197, 198, 201, 202, 204, 209, 210, 212, 216, 225, 226, 228, 232, 240 };
  125. for(size_t i=0;i<5;i++){
  126. uint8_t ok;
  127. do{
  128. ok=1; secret[i]=0;
  129. for(size_t j=0;j<64;j+=8) secret[i]|=((uint64_t)c[wyrand(&seed)%sizeof(c)])<<j;
  130. if(secret[i]%2==0){ ok=0; continue; }
  131. for(size_t j=0;j<i;j++)
  132. #if defined(__GNUC__) || defined(__INTEL_COMPILER) || defined(__clang__)
  133. if(__builtin_popcountll(secret[j]^secret[i])!=32){ ok=0; break; }
  134. #elif defined(_MSC_VER) && defined(_M_X64)
  135. if(_mm_popcnt_u64(secret[j]^secret[i])!=32){ ok=0; break; }
  136. #endif
  137. if(!ok)continue;
  138. for(uint64_t j=3;j<0x100000000ull;j+=2) if(secret[i]%j==0){ ok=0; break; }
  139. }while(!ok);
  140. }
  141. }
  142. #endif