Radish alpha
H
rad:z3QDZAW2FAfuLvihrhiyDC9fAD8G9
HardenedBSD Package Manager
Radicle
Git
cleanup: move mum.h into external/include
Baptiste Daroussin committed 4 years ago
commit 128504d1fd71ae97b3fd36b2e487fe2f717e4142
parent 4dcf962
3 files changed +422 -423
added external/include/mum.h
@@ -0,0 +1,422 @@
+
/* Copyright (c) 2016, 2017, 2018
+
   Vladimir Makarov <vmakarov@gcc.gnu.org>
+

+
   Permission is hereby granted, free of charge, to any person
+
   obtaining a copy of this software and associated documentation
+
   files (the "Software"), to deal in the Software without
+
   restriction, including without limitation the rights to use, copy,
+
   modify, merge, publish, distribute, sublicense, and/or sell copies
+
   of the Software, and to permit persons to whom the Software is
+
   furnished to do so, subject to the following conditions:
+

+
   The above copyright notice and this permission notice shall be
+
   included in all copies or substantial portions of the Software.
+

+
   THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+
   EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+
   MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+
   NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
+
   BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
+
   ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
+
   CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+
   SOFTWARE.
+
*/
+

+
/* This file implements MUM (MUltiply and Mix) hashing.  We randomize
+
   input data by 64x64-bit multiplication and mixing hi- and low-parts
+
   of the multiplication result by using an addition and then mix it
+
   into the current state.  We use prime numbers randomly generated
+
   with the equal probability of their bit values for the
+
   multiplication.  When all primes are used once, the state is
+
   randomized and the same prime numbers are used again for data
+
   randomization.
+

+
   The MUM hashing passes all SMHasher tests.  Pseudo Random Number
+
   Generator based on MUM also passes NIST Statistical Test Suite for
+
   Random and Pseudorandom Number Generators for Cryptographic
+
   Applications (version 2.2.1) with 1000 bitstreams each containing
+
   1M bits.  MUM hashing is also faster Spooky64 and City64 on small
+
   strings (at least upto 512-bit) on Haswell and Power7.  The MUM bulk
+
   speed (speed on very long data) is bigger than Spooky and City on
+
   Power7.  On Haswell the bulk speed is bigger than Spooky one and
+
   close to City speed.  */
+

+
#ifndef __MUM_HASH__
+
#define __MUM_HASH__
+

+
#include <stddef.h>
+
#include <stdlib.h>
+
#include <string.h>
+
#include <limits.h>
+

+
#ifdef _MSC_VER
+
typedef unsigned __int16 uint16_t;
+
typedef unsigned __int32 uint32_t;
+
typedef unsigned __int64 uint64_t;
+
#else
+
#include <stdint.h>
+
#endif
+

+
#ifdef __GNUC__
+
#define _MUM_ATTRIBUTE_UNUSED  __attribute__((unused))
+
#define _MUM_OPTIMIZE(opts) __attribute__((__optimize__ (opts)))
+
#define _MUM_TARGET(opts) __attribute__((__target__ (opts)))
+
#else
+
#define _MUM_ATTRIBUTE_UNUSED
+
#define _MUM_OPTIMIZE(opts)
+
#define _MUM_TARGET(opts)
+
#endif
+

+
/* Macro saying to use 128-bit integers implemented by GCC for some
+
   targets.  */
+
#ifndef _MUM_USE_INT128
+
/* In GCC uint128_t is defined if HOST_BITS_PER_WIDE_INT >= 64.
+
   HOST_WIDE_INT is long if HOST_BITS_PER_LONG > HOST_BITS_PER_INT,
+
   otherwise int. */
+
#if defined(__GNUC__) && UINT_MAX != ULONG_MAX
+
#define _MUM_USE_INT128 1
+
#else
+
#define _MUM_USE_INT128 0
+
#endif
+
#endif
+

+
/* Here are different primes randomly generated with the equal
+
   probability of their bit values.  They are used to randomize input
+
   values.  */
+
static uint64_t _mum_hash_step_prime = 0x2e0bb864e9ea7df5ULL;
+
static uint64_t _mum_key_step_prime = 0xcdb32970830fcaa1ULL;
+
static uint64_t _mum_block_start_prime = 0xc42b5e2e6480b23bULL;
+
static uint64_t _mum_unroll_prime = 0x7b51ec3d22f7096fULL;
+
static uint64_t _mum_tail_prime = 0xaf47d47c99b1461bULL;
+
static uint64_t _mum_finish_prime1 = 0xa9a7ae7ceff79f3fULL;
+
static uint64_t _mum_finish_prime2 = 0xaf47d47c99b1461bULL;
+
  
+
static uint64_t _mum_primes [] = {
+
  0X9ebdcae10d981691, 0X32b9b9b97a27ac7d, 0X29b5584d83d35bbd, 0X4b04e0e61401255f,
+
  0X25e8f7b1f1c9d027, 0X80d4c8c000f3e881, 0Xbd1255431904b9dd, 0X8a3bd4485eee6d81,
+
  0X3bc721b2aad05197, 0X71b1a19b907d6e33, 0X525e6c1084a8534b, 0X9e4c2cd340c1299f,
+
  0Xde3add92e94caa37, 0X7e14eadb1f65311d, 0X3f5aa40f89812853, 0X33b15a3b587d15c9,
+
};
+

+
/* Multiply 64-bit V and P and return sum of high and low parts of the
+
   result.  */
+
static inline uint64_t
+
_mum (uint64_t v, uint64_t p) {
+
  uint64_t hi, lo;
+
#if _MUM_USE_INT128
+
#if defined(__aarch64__)
+
  /* AARCH64 needs 2 insns to calculate 128-bit result of the
+
     multiplication.  If we use a generic code we actually call a
+
     function doing 128x128->128 bit multiplication.  The function is
+
     very slow.  */
+
  lo = v * p;
+
  asm ("umulh %0, %1, %2" : "=r" (hi) : "r" (v), "r" (p));
+
#else
+
  __uint128_t r = (__uint128_t) v * (__uint128_t) p;
+
  hi = (uint64_t) (r >> 64);
+
  lo = (uint64_t) r;
+
#endif
+
#else
+
  /* Implementation of 64x64->128-bit multiplication by four 32x32->64
+
     bit multiplication.  */
+
  uint64_t hv = v >> 32, hp = p >> 32;
+
  uint64_t lv = (uint32_t) v, lp = (uint32_t) p;
+
  uint64_t rh =  hv * hp;
+
  uint64_t rm_0 = hv * lp;
+
  uint64_t rm_1 = hp * lv;
+
  uint64_t rl =  lv * lp;
+
  uint64_t t, carry = 0;
+
  
+
  /* We could ignore a carry bit here if we did not care about the
+
     same hash for 32-bit and 64-bit targets.  */
+
  t = rl + (rm_0 << 32);
+
#ifdef MUM_TARGET_INDEPENDENT_HASH
+
  carry = t < rl;
+
#endif
+
  lo = t + (rm_1 << 32);
+
#ifdef MUM_TARGET_INDEPENDENT_HASH
+
  carry += lo < t;
+
#endif
+
  hi = rh + (rm_0 >> 32) + (rm_1 >> 32) + carry;
+
#endif
+
  /* We could use XOR here too but, for some reasons, on Haswell and
+
     Power7 using an addition improves hashing performance by 10% for
+
     small strings.  */
+
  return hi + lo;
+
}
+

+
#if defined(_MSC_VER)
+
#define _mum_bswap_32(x) _byteswap_uint32_t (x)
+
#define _mum_bswap_64(x) _byteswap_uint64_t (x)
+
#elif defined(__APPLE__)
+
#include <libkern/OSByteOrder.h>
+
#define _mum_bswap_32(x) OSSwapInt32 (x)
+
#define _mum_bswap_64(x) OSSwapInt64 (x)
+
#elif defined(__GNUC__)
+
#define _mum_bswap32(x) __builtin_bswap32 (x)
+
#define _mum_bswap64(x) __builtin_bswap64 (x)
+
#else
+
#include <byteswap.h>
+
#define _mum_bswap32(x) bswap32 (x)
+
#define _mum_bswap64(x) bswap64 (x)
+
#endif
+

+
static inline uint64_t
+
_mum_le (uint64_t v) {
+
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
+
  return v;
+
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
+
  return _mum_bswap64 (v);
+
#else
+
#error "Unknown endianess"
+
#endif
+
}
+

+
static inline uint32_t
+
_mum_le32 (uint32_t v) {
+
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
+
  return v;
+
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
+
  return _mum_bswap32 (v);
+
#else
+
#error "Unknown endianess"
+
#endif
+
}
+

+
static inline uint64_t
+
_mum_le16 (uint16_t v) {
+
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
+
  return v;
+
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
+
  return (v >> 8) | ((v & 0xff) << 8);
+
#else
+
#error "Unknown endianess"
+
#endif
+
}
+

+
/* Macro defining how many times the most nested loop in
+
   _mum_hash_aligned will be unrolled by the compiler (although it can
+
   make an own decision:).  Use only a constant here to help a
+
   compiler to unroll a major loop.
+

+
   The macro value affects the result hash for strings > 128 bit.  The
+
   unroll factor greatly affects the hashing speed.  We prefer the
+
   speed.  */
+
#ifndef _MUM_UNROLL_FACTOR_POWER
+
#if defined(__PPC64__) && !defined(MUM_TARGET_INDEPENDENT_HASH)
+
#define _MUM_UNROLL_FACTOR_POWER 3
+
#elif defined(__aarch64__) && !defined(MUM_TARGET_INDEPENDENT_HASH)
+
#define _MUM_UNROLL_FACTOR_POWER 4
+
#elif defined (MUM_V1) || defined (MUM_V2)
+
#define _MUM_UNROLL_FACTOR_POWER 2
+
#else
+
#define _MUM_UNROLL_FACTOR_POWER 3
+
#endif
+
#endif
+

+
#if _MUM_UNROLL_FACTOR_POWER < 1
+
#error "too small unroll factor"
+
#elif _MUM_UNROLL_FACTOR_POWER > 4
+
#error "We have not enough primes for such unroll factor"
+
#endif
+

+
#define _MUM_UNROLL_FACTOR (1 << _MUM_UNROLL_FACTOR_POWER)
+

+
/* Rotate V left by SH.  */
+
static inline uint64_t _mum_rotl (uint64_t v, int sh) {
+
  return v << sh | v >> (64 - sh);
+
}
+

+
static inline uint64_t _MUM_OPTIMIZE("unroll-loops")
+
_mum_hash_aligned (uint64_t start, const void *key, size_t len) {
+
  uint64_t result = start;
+
  const unsigned char *str = (const unsigned char *) key;
+
  uint64_t u64;
+
  size_t i;
+
  size_t n;
+
  
+
#ifndef MUM_V2
+
  result = _mum (result, _mum_block_start_prime);
+
#endif
+
  while  (len > _MUM_UNROLL_FACTOR * sizeof (uint64_t)) {
+
    /* This loop could be vectorized when we have vector insns for
+
       64x64->128-bit multiplication.  AVX2 currently only have vector
+
       insns for 4 32x32->64-bit multiplication and for 1
+
       64x64->128-bit multiplication (pclmulqdq).  */
+
#if defined (MUM_V1) || defined (MUM_V2)
+
    for (i = 0; i < _MUM_UNROLL_FACTOR; i++)
+
      result ^= _mum (_mum_le (((uint64_t *) str)[i]), _mum_primes[i]);
+
#else
+
    for (i = 0; i < _MUM_UNROLL_FACTOR; i += 2)
+
      result ^= _mum (_mum_le (((uint64_t *) str)[i]) ^ _mum_primes[i],
+
		      _mum_le (((uint64_t *) str)[i + 1]) ^ _mum_primes[i + 1]);
+
#endif
+
    len -= _MUM_UNROLL_FACTOR * sizeof (uint64_t);
+
    str += _MUM_UNROLL_FACTOR * sizeof (uint64_t);
+
    /* We will use the same prime numbers on the next iterations --
+
       randomize the state.  */
+
    result = _mum (result, _mum_unroll_prime);
+
  }
+
  n = len / sizeof (uint64_t);
+
  for (i = 0; i < n; i++)
+
    result ^= _mum (_mum_le (((uint64_t *) str)[i]), _mum_primes[i]);
+
  len -= n * sizeof (uint64_t); str += n * sizeof (uint64_t);
+
  switch (len) {
+
  case 7:
+
    u64 = _mum_le32 (*(uint32_t *) str);
+
    u64 |= _mum_le16 (*(uint16_t *) (str + 4)) << 32;
+
    u64 |= (uint64_t) str[6] << 48;
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 6:
+
    u64 = _mum_le32 (*(uint32_t *) str);
+
    u64 |= _mum_le16 (*(uint16_t *) (str + 4)) << 32;
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 5:
+
    u64 = _mum_le32 (*(uint32_t *) str);
+
    u64 |= (uint64_t) str[4] << 32;
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 4:
+
    u64 = _mum_le32 (*(uint32_t *) str);
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 3:
+
    u64 = _mum_le16 (*(uint16_t *) str);
+
    u64 |= (uint64_t) str[2] << 16;
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 2:
+
    u64 = _mum_le16 (*(uint16_t *) str);
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  case 1:
+
    u64 = str[0];
+
    return result ^ _mum (u64, _mum_tail_prime);
+
  }
+
  return result;
+
}
+

+
/* Final randomization of H.  */
+
static inline uint64_t
+
_mum_final (uint64_t h) {
+
#if defined (MUM_V1)
+
  h ^= _mum (h, _mum_finish_prime1);
+
  h ^= _mum (h, _mum_finish_prime2);
+
#elif defined (MUM_V2)
+
  h ^= _mum_rotl (h, 33);
+
  h ^= _mum (h, _mum_finish_prime1);
+
#else
+
  h = _mum (h, h);
+
#endif
+
  return h;
+
}
+

+
#ifndef _MUM_UNALIGNED_ACCESS
+
#if defined(__x86_64__) || defined(__i386__) || defined(__PPC64__) \
+
    || defined(__s390__) || defined(__m32c__) || defined(cris)     \
+
    || defined(__CR16__) || defined(__vax__) || defined(__m68k__) \
+
    || defined(__aarch64__) || defined(_M_AMD64) || defined(_M_IX86)
+
#define _MUM_UNALIGNED_ACCESS 1
+
#else
+
#define _MUM_UNALIGNED_ACCESS 0
+
#endif
+
#endif
+

+
/* When we need an aligned access to data being hashed we move part of
+
   the unaligned data to an aligned block of given size and then
+
   process it, repeating processing the data by the block.  */
+
#ifndef _MUM_BLOCK_LEN
+
#define _MUM_BLOCK_LEN 1024
+
#endif
+

+
#if _MUM_BLOCK_LEN < 8
+
#error "too small block length"
+
#endif
+

+
static inline uint64_t
+
#if defined(__x86_64__)
+
_MUM_TARGET("inline-all-stringops")
+
#endif
+
_mum_hash_default (const void *key, size_t len, uint64_t seed) {
+
  uint64_t result;
+
  const unsigned char *str = (const unsigned char *) key;
+
  size_t block_len;
+
  uint64_t buf[_MUM_BLOCK_LEN / sizeof (uint64_t)];
+
  
+
  result = seed + len;
+
  if (((size_t) str & 0x7) == 0)
+
    result = _mum_hash_aligned (result, key, len);
+
  else {
+
    while (len != 0) {
+
      block_len = len < _MUM_BLOCK_LEN ? len : _MUM_BLOCK_LEN;
+
      memmove (buf, str, block_len);
+
      result = _mum_hash_aligned (result, buf, block_len);
+
      len -= block_len;
+
      str += block_len;
+
    }
+
  }
+
  return _mum_final (result);
+
}
+

+
static inline uint64_t
+
_mum_next_factor (void) {
+
  uint64_t start = 0;
+
  int i;
+
  
+
  for (i = 0; i < 8; i++)
+
    start = (start << 8) | rand() % 256;
+
  return start;
+
}
+

+
/* ++++++++++++++++++++++++++ Interface functions: +++++++++++++++++++  */
+

+
/* Set random multiplicators depending on SEED.  */
+
static inline void
+
mum_hash_randomize (uint64_t seed) {
+
  size_t i;
+

+
  srand (seed);
+
  _mum_hash_step_prime = _mum_next_factor ();
+
  _mum_key_step_prime = _mum_next_factor ();
+
  _mum_finish_prime1 = _mum_next_factor ();
+
  _mum_finish_prime2 = _mum_next_factor ();
+
  _mum_block_start_prime = _mum_next_factor ();
+
  _mum_unroll_prime = _mum_next_factor ();
+
  _mum_tail_prime = _mum_next_factor ();
+
  for (i = 0; i < sizeof (_mum_primes) / sizeof (uint64_t); i++)
+
    _mum_primes[i] = _mum_next_factor ();
+
}
+

+
/* Start hashing data with SEED.  Return the state.  */
+
static inline uint64_t
+
mum_hash_init (uint64_t seed) {
+
  return seed;
+
}
+

+
/* Process data KEY with the state H and return the updated state.  */
+
static inline uint64_t
+
mum_hash_step (uint64_t h, uint64_t key) {
+
  return _mum (h, _mum_hash_step_prime) ^ _mum (key, _mum_key_step_prime);
+
}
+

+
/* Return the result of hashing using the current state H.  */
+
static inline uint64_t
+
mum_hash_finish (uint64_t h) {
+
  return _mum_final (h);
+
}
+

+
/* Fast hashing of KEY with SEED.  The hash is always the same for the
+
   same key on any target. */
+
static inline size_t
+
mum_hash64 (uint64_t key, uint64_t seed) {
+
  return mum_hash_finish (mum_hash_step (mum_hash_init (seed), key));
+
}
+

+
/* Hash data KEY of length LEN and SEED.  The hash depends on the
+
   target endianess and the unroll factor.  */
+
static inline uint64_t
+
mum_hash (const void *key, size_t len, uint64_t seed) {
+
#if _MUM_UNALIGNED_ACCESS
+
  return _mum_final (_mum_hash_aligned (seed + len, key, len));
+
#else
+
  return _mum_hash_default (key, len, seed);
+
#endif
+
}
+

+
#endif
deleted external/mum/mum.h
@@ -1,422 +0,0 @@
-
/* Copyright (c) 2016, 2017, 2018
-
   Vladimir Makarov <vmakarov@gcc.gnu.org>
-

-
   Permission is hereby granted, free of charge, to any person
-
   obtaining a copy of this software and associated documentation
-
   files (the "Software"), to deal in the Software without
-
   restriction, including without limitation the rights to use, copy,
-
   modify, merge, publish, distribute, sublicense, and/or sell copies
-
   of the Software, and to permit persons to whom the Software is
-
   furnished to do so, subject to the following conditions:
-

-
   The above copyright notice and this permission notice shall be
-
   included in all copies or substantial portions of the Software.
-

-
   THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
-
   EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
-
   MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
-
   NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
-
   BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
-
   ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
-
   CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
-
   SOFTWARE.
-
*/
-

-
/* This file implements MUM (MUltiply and Mix) hashing.  We randomize
-
   input data by 64x64-bit multiplication and mixing hi- and low-parts
-
   of the multiplication result by using an addition and then mix it
-
   into the current state.  We use prime numbers randomly generated
-
   with the equal probability of their bit values for the
-
   multiplication.  When all primes are used once, the state is
-
   randomized and the same prime numbers are used again for data
-
   randomization.
-

-
   The MUM hashing passes all SMHasher tests.  Pseudo Random Number
-
   Generator based on MUM also passes NIST Statistical Test Suite for
-
   Random and Pseudorandom Number Generators for Cryptographic
-
   Applications (version 2.2.1) with 1000 bitstreams each containing
-
   1M bits.  MUM hashing is also faster Spooky64 and City64 on small
-
   strings (at least upto 512-bit) on Haswell and Power7.  The MUM bulk
-
   speed (speed on very long data) is bigger than Spooky and City on
-
   Power7.  On Haswell the bulk speed is bigger than Spooky one and
-
   close to City speed.  */
-

-
#ifndef __MUM_HASH__
-
#define __MUM_HASH__
-

-
#include <stddef.h>
-
#include <stdlib.h>
-
#include <string.h>
-
#include <limits.h>
-

-
#ifdef _MSC_VER
-
typedef unsigned __int16 uint16_t;
-
typedef unsigned __int32 uint32_t;
-
typedef unsigned __int64 uint64_t;
-
#else
-
#include <stdint.h>
-
#endif
-

-
#ifdef __GNUC__
-
#define _MUM_ATTRIBUTE_UNUSED  __attribute__((unused))
-
#define _MUM_OPTIMIZE(opts) __attribute__((__optimize__ (opts)))
-
#define _MUM_TARGET(opts) __attribute__((__target__ (opts)))
-
#else
-
#define _MUM_ATTRIBUTE_UNUSED
-
#define _MUM_OPTIMIZE(opts)
-
#define _MUM_TARGET(opts)
-
#endif
-

-
/* Macro saying to use 128-bit integers implemented by GCC for some
-
   targets.  */
-
#ifndef _MUM_USE_INT128
-
/* In GCC uint128_t is defined if HOST_BITS_PER_WIDE_INT >= 64.
-
   HOST_WIDE_INT is long if HOST_BITS_PER_LONG > HOST_BITS_PER_INT,
-
   otherwise int. */
-
#if defined(__GNUC__) && UINT_MAX != ULONG_MAX
-
#define _MUM_USE_INT128 1
-
#else
-
#define _MUM_USE_INT128 0
-
#endif
-
#endif
-

-
/* Here are different primes randomly generated with the equal
-
   probability of their bit values.  They are used to randomize input
-
   values.  */
-
static uint64_t _mum_hash_step_prime = 0x2e0bb864e9ea7df5ULL;
-
static uint64_t _mum_key_step_prime = 0xcdb32970830fcaa1ULL;
-
static uint64_t _mum_block_start_prime = 0xc42b5e2e6480b23bULL;
-
static uint64_t _mum_unroll_prime = 0x7b51ec3d22f7096fULL;
-
static uint64_t _mum_tail_prime = 0xaf47d47c99b1461bULL;
-
static uint64_t _mum_finish_prime1 = 0xa9a7ae7ceff79f3fULL;
-
static uint64_t _mum_finish_prime2 = 0xaf47d47c99b1461bULL;
-
  
-
static uint64_t _mum_primes [] = {
-
  0X9ebdcae10d981691, 0X32b9b9b97a27ac7d, 0X29b5584d83d35bbd, 0X4b04e0e61401255f,
-
  0X25e8f7b1f1c9d027, 0X80d4c8c000f3e881, 0Xbd1255431904b9dd, 0X8a3bd4485eee6d81,
-
  0X3bc721b2aad05197, 0X71b1a19b907d6e33, 0X525e6c1084a8534b, 0X9e4c2cd340c1299f,
-
  0Xde3add92e94caa37, 0X7e14eadb1f65311d, 0X3f5aa40f89812853, 0X33b15a3b587d15c9,
-
};
-

-
/* Multiply 64-bit V and P and return sum of high and low parts of the
-
   result.  */
-
static inline uint64_t
-
_mum (uint64_t v, uint64_t p) {
-
  uint64_t hi, lo;
-
#if _MUM_USE_INT128
-
#if defined(__aarch64__)
-
  /* AARCH64 needs 2 insns to calculate 128-bit result of the
-
     multiplication.  If we use a generic code we actually call a
-
     function doing 128x128->128 bit multiplication.  The function is
-
     very slow.  */
-
  lo = v * p;
-
  asm ("umulh %0, %1, %2" : "=r" (hi) : "r" (v), "r" (p));
-
#else
-
  __uint128_t r = (__uint128_t) v * (__uint128_t) p;
-
  hi = (uint64_t) (r >> 64);
-
  lo = (uint64_t) r;
-
#endif
-
#else
-
  /* Implementation of 64x64->128-bit multiplication by four 32x32->64
-
     bit multiplication.  */
-
  uint64_t hv = v >> 32, hp = p >> 32;
-
  uint64_t lv = (uint32_t) v, lp = (uint32_t) p;
-
  uint64_t rh =  hv * hp;
-
  uint64_t rm_0 = hv * lp;
-
  uint64_t rm_1 = hp * lv;
-
  uint64_t rl =  lv * lp;
-
  uint64_t t, carry = 0;
-
  
-
  /* We could ignore a carry bit here if we did not care about the
-
     same hash for 32-bit and 64-bit targets.  */
-
  t = rl + (rm_0 << 32);
-
#ifdef MUM_TARGET_INDEPENDENT_HASH
-
  carry = t < rl;
-
#endif
-
  lo = t + (rm_1 << 32);
-
#ifdef MUM_TARGET_INDEPENDENT_HASH
-
  carry += lo < t;
-
#endif
-
  hi = rh + (rm_0 >> 32) + (rm_1 >> 32) + carry;
-
#endif
-
  /* We could use XOR here too but, for some reasons, on Haswell and
-
     Power7 using an addition improves hashing performance by 10% for
-
     small strings.  */
-
  return hi + lo;
-
}
-

-
#if defined(_MSC_VER)
-
#define _mum_bswap_32(x) _byteswap_uint32_t (x)
-
#define _mum_bswap_64(x) _byteswap_uint64_t (x)
-
#elif defined(__APPLE__)
-
#include <libkern/OSByteOrder.h>
-
#define _mum_bswap_32(x) OSSwapInt32 (x)
-
#define _mum_bswap_64(x) OSSwapInt64 (x)
-
#elif defined(__GNUC__)
-
#define _mum_bswap32(x) __builtin_bswap32 (x)
-
#define _mum_bswap64(x) __builtin_bswap64 (x)
-
#else
-
#include <byteswap.h>
-
#define _mum_bswap32(x) bswap32 (x)
-
#define _mum_bswap64(x) bswap64 (x)
-
#endif
-

-
static inline uint64_t
-
_mum_le (uint64_t v) {
-
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
-
  return v;
-
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-
  return _mum_bswap64 (v);
-
#else
-
#error "Unknown endianess"
-
#endif
-
}
-

-
static inline uint32_t
-
_mum_le32 (uint32_t v) {
-
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
-
  return v;
-
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-
  return _mum_bswap32 (v);
-
#else
-
#error "Unknown endianess"
-
#endif
-
}
-

-
static inline uint64_t
-
_mum_le16 (uint16_t v) {
-
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ || !defined(MUM_TARGET_INDEPENDENT_HASH)
-
  return v;
-
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
-
  return (v >> 8) | ((v & 0xff) << 8);
-
#else
-
#error "Unknown endianess"
-
#endif
-
}
-

-
/* Macro defining how many times the most nested loop in
-
   _mum_hash_aligned will be unrolled by the compiler (although it can
-
   make an own decision:).  Use only a constant here to help a
-
   compiler to unroll a major loop.
-

-
   The macro value affects the result hash for strings > 128 bit.  The
-
   unroll factor greatly affects the hashing speed.  We prefer the
-
   speed.  */
-
#ifndef _MUM_UNROLL_FACTOR_POWER
-
#if defined(__PPC64__) && !defined(MUM_TARGET_INDEPENDENT_HASH)
-
#define _MUM_UNROLL_FACTOR_POWER 3
-
#elif defined(__aarch64__) && !defined(MUM_TARGET_INDEPENDENT_HASH)
-
#define _MUM_UNROLL_FACTOR_POWER 4
-
#elif defined (MUM_V1) || defined (MUM_V2)
-
#define _MUM_UNROLL_FACTOR_POWER 2
-
#else
-
#define _MUM_UNROLL_FACTOR_POWER 3
-
#endif
-
#endif
-

-
#if _MUM_UNROLL_FACTOR_POWER < 1
-
#error "too small unroll factor"
-
#elif _MUM_UNROLL_FACTOR_POWER > 4
-
#error "We have not enough primes for such unroll factor"
-
#endif
-

-
#define _MUM_UNROLL_FACTOR (1 << _MUM_UNROLL_FACTOR_POWER)
-

-
/* Rotate V left by SH.  */
-
static inline uint64_t _mum_rotl (uint64_t v, int sh) {
-
  return v << sh | v >> (64 - sh);
-
}
-

-
static inline uint64_t _MUM_OPTIMIZE("unroll-loops")
-
_mum_hash_aligned (uint64_t start, const void *key, size_t len) {
-
  uint64_t result = start;
-
  const unsigned char *str = (const unsigned char *) key;
-
  uint64_t u64;
-
  size_t i;
-
  size_t n;
-
  
-
#ifndef MUM_V2
-
  result = _mum (result, _mum_block_start_prime);
-
#endif
-
  while  (len > _MUM_UNROLL_FACTOR * sizeof (uint64_t)) {
-
    /* This loop could be vectorized when we have vector insns for
-
       64x64->128-bit multiplication.  AVX2 currently only have vector
-
       insns for 4 32x32->64-bit multiplication and for 1
-
       64x64->128-bit multiplication (pclmulqdq).  */
-
#if defined (MUM_V1) || defined (MUM_V2)
-
    for (i = 0; i < _MUM_UNROLL_FACTOR; i++)
-
      result ^= _mum (_mum_le (((uint64_t *) str)[i]), _mum_primes[i]);
-
#else
-
    for (i = 0; i < _MUM_UNROLL_FACTOR; i += 2)
-
      result ^= _mum (_mum_le (((uint64_t *) str)[i]) ^ _mum_primes[i],
-
		      _mum_le (((uint64_t *) str)[i + 1]) ^ _mum_primes[i + 1]);
-
#endif
-
    len -= _MUM_UNROLL_FACTOR * sizeof (uint64_t);
-
    str += _MUM_UNROLL_FACTOR * sizeof (uint64_t);
-
    /* We will use the same prime numbers on the next iterations --
-
       randomize the state.  */
-
    result = _mum (result, _mum_unroll_prime);
-
  }
-
  n = len / sizeof (uint64_t);
-
  for (i = 0; i < n; i++)
-
    result ^= _mum (_mum_le (((uint64_t *) str)[i]), _mum_primes[i]);
-
  len -= n * sizeof (uint64_t); str += n * sizeof (uint64_t);
-
  switch (len) {
-
  case 7:
-
    u64 = _mum_le32 (*(uint32_t *) str);
-
    u64 |= _mum_le16 (*(uint16_t *) (str + 4)) << 32;
-
    u64 |= (uint64_t) str[6] << 48;
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 6:
-
    u64 = _mum_le32 (*(uint32_t *) str);
-
    u64 |= _mum_le16 (*(uint16_t *) (str + 4)) << 32;
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 5:
-
    u64 = _mum_le32 (*(uint32_t *) str);
-
    u64 |= (uint64_t) str[4] << 32;
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 4:
-
    u64 = _mum_le32 (*(uint32_t *) str);
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 3:
-
    u64 = _mum_le16 (*(uint16_t *) str);
-
    u64 |= (uint64_t) str[2] << 16;
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 2:
-
    u64 = _mum_le16 (*(uint16_t *) str);
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  case 1:
-
    u64 = str[0];
-
    return result ^ _mum (u64, _mum_tail_prime);
-
  }
-
  return result;
-
}
-

-
/* Final randomization of H.  */
-
static inline uint64_t
-
_mum_final (uint64_t h) {
-
#if defined (MUM_V1)
-
  h ^= _mum (h, _mum_finish_prime1);
-
  h ^= _mum (h, _mum_finish_prime2);
-
#elif defined (MUM_V2)
-
  h ^= _mum_rotl (h, 33);
-
  h ^= _mum (h, _mum_finish_prime1);
-
#else
-
  h = _mum (h, h);
-
#endif
-
  return h;
-
}
-

-
#ifndef _MUM_UNALIGNED_ACCESS
-
#if defined(__x86_64__) || defined(__i386__) || defined(__PPC64__) \
-
    || defined(__s390__) || defined(__m32c__) || defined(cris)     \
-
    || defined(__CR16__) || defined(__vax__) || defined(__m68k__) \
-
    || defined(__aarch64__) || defined(_M_AMD64) || defined(_M_IX86)
-
#define _MUM_UNALIGNED_ACCESS 1
-
#else
-
#define _MUM_UNALIGNED_ACCESS 0
-
#endif
-
#endif
-

-
/* When we need an aligned access to data being hashed we move part of
-
   the unaligned data to an aligned block of given size and then
-
   process it, repeating processing the data by the block.  */
-
#ifndef _MUM_BLOCK_LEN
-
#define _MUM_BLOCK_LEN 1024
-
#endif
-

-
#if _MUM_BLOCK_LEN < 8
-
#error "too small block length"
-
#endif
-

-
static inline uint64_t
-
#if defined(__x86_64__)
-
_MUM_TARGET("inline-all-stringops")
-
#endif
-
_mum_hash_default (const void *key, size_t len, uint64_t seed) {
-
  uint64_t result;
-
  const unsigned char *str = (const unsigned char *) key;
-
  size_t block_len;
-
  uint64_t buf[_MUM_BLOCK_LEN / sizeof (uint64_t)];
-
  
-
  result = seed + len;
-
  if (((size_t) str & 0x7) == 0)
-
    result = _mum_hash_aligned (result, key, len);
-
  else {
-
    while (len != 0) {
-
      block_len = len < _MUM_BLOCK_LEN ? len : _MUM_BLOCK_LEN;
-
      memmove (buf, str, block_len);
-
      result = _mum_hash_aligned (result, buf, block_len);
-
      len -= block_len;
-
      str += block_len;
-
    }
-
  }
-
  return _mum_final (result);
-
}
-

-
static inline uint64_t
-
_mum_next_factor (void) {
-
  uint64_t start = 0;
-
  int i;
-
  
-
  for (i = 0; i < 8; i++)
-
    start = (start << 8) | rand() % 256;
-
  return start;
-
}
-

-
/* ++++++++++++++++++++++++++ Interface functions: +++++++++++++++++++  */
-

-
/* Set random multiplicators depending on SEED.  */
-
static inline void
-
mum_hash_randomize (uint64_t seed) {
-
  size_t i;
-

-
  srand (seed);
-
  _mum_hash_step_prime = _mum_next_factor ();
-
  _mum_key_step_prime = _mum_next_factor ();
-
  _mum_finish_prime1 = _mum_next_factor ();
-
  _mum_finish_prime2 = _mum_next_factor ();
-
  _mum_block_start_prime = _mum_next_factor ();
-
  _mum_unroll_prime = _mum_next_factor ();
-
  _mum_tail_prime = _mum_next_factor ();
-
  for (i = 0; i < sizeof (_mum_primes) / sizeof (uint64_t); i++)
-
    _mum_primes[i] = _mum_next_factor ();
-
}
-

-
/* Start hashing data with SEED.  Return the state.  */
-
static inline uint64_t
-
mum_hash_init (uint64_t seed) {
-
  return seed;
-
}
-

-
/* Process data KEY with the state H and return the updated state.  */
-
static inline uint64_t
-
mum_hash_step (uint64_t h, uint64_t key) {
-
  return _mum (h, _mum_hash_step_prime) ^ _mum (key, _mum_key_step_prime);
-
}
-

-
/* Return the result of hashing using the current state H.  */
-
static inline uint64_t
-
mum_hash_finish (uint64_t h) {
-
  return _mum_final (h);
-
}
-

-
/* Fast hashing of KEY with SEED.  The hash is always the same for the
-
   same key on any target. */
-
static inline size_t
-
mum_hash64 (uint64_t key, uint64_t seed) {
-
  return mum_hash_finish (mum_hash_step (mum_hash_init (seed), key));
-
}
-

-
/* Hash data KEY of length LEN and SEED.  The hash depends on the
-
   target endianess and the unroll factor.  */
-
static inline uint64_t
-
mum_hash (const void *key, size_t len, uint64_t seed) {
-
#if _MUM_UNALIGNED_ACCESS
-
  return _mum_final (_mum_hash_aligned (seed + len, key, len));
-
#else
-
  return _mum_hash_default (key, len, seed);
-
#endif
-
}
-

-
#endif
modified libpkg/Makefile.autosetup
@@ -57,7 +57,6 @@ LOCAL_CFLAGS= -I$(top_srcdir)/compat \
		-I$(top_srcdir)/external/picosat \
		-I$(top_srcdir)/external/yxml \
		-I$(top_srcdir)/external/include \
-
		-I$(top_srcdir)/external/mum \
		-I$(top_srcdir)/external/libucl/include \
		-I$(top_srcdir)/external/libfetch \
		-I$(top_srcdir)/external/lua/src \