Template instances can allocate BIHASH_KVP_PER_PAGE data records tangent to the bucket, to remove a dependent read / prefetch. Template instances can ask for immediate memory allocation, to avoid several branches in the lookup path. Clean up l2 fib, gpb plugin codes: use clib_bihash_get_bucket(...) Use hugepages for bihash allocation arenas Type: improvement Signed-off-by: Dave Barach <dave@barachs.net> Signed-off-by: Damjan Marion <damarion@cisco.com> Change-Id: I92fc11bc58e48d84e2d61f44580916dd1c56361c
113 lines
3.1 KiB
C
113 lines
3.1 KiB
C
/*
|
|
* Copyright (c) 2016 Cisco and/or its affiliates.
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#undef BIHASH_TYPE
|
|
#undef BIHASH_KVP_PER_PAGE
|
|
#undef BIHASH_32_64_SVM
|
|
#undef BIHASH_ENABLE_STATS
|
|
#undef BIHASH_KVP_AT_BUCKET_LEVEL
|
|
#undef BIHASH_LAZY_INSTANTIATE
|
|
#undef BIHASH_BUCKET_PREFETCH_CACHE_LINES
|
|
|
|
#define BIHASH_TYPE _48_8
|
|
#define BIHASH_KVP_PER_PAGE 4
|
|
#define BIHASH_KVP_AT_BUCKET_LEVEL 0
|
|
#define BIHASH_LAZY_INSTANTIATE 1
|
|
#define BIHASH_BUCKET_PREFETCH_CACHE_LINES 1
|
|
|
|
#ifndef __included_bihash_48_8_h__
|
|
#define __included_bihash_48_8_h__
|
|
|
|
#include <vppinfra/crc32.h>
|
|
#include <vppinfra/heap.h>
|
|
#include <vppinfra/format.h>
|
|
#include <vppinfra/pool.h>
|
|
#include <vppinfra/xxhash.h>
|
|
|
|
typedef struct
|
|
{
|
|
u64 key[6];
|
|
u64 value;
|
|
} clib_bihash_kv_48_8_t;
|
|
|
|
static inline int
|
|
clib_bihash_is_free_48_8 (const clib_bihash_kv_48_8_t * v)
|
|
{
|
|
/* Free values are clib_memset to 0xff, check a bit... */
|
|
if (v->key[0] == ~0ULL && v->value == ~0ULL)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
static inline u64
|
|
clib_bihash_hash_48_8 (const clib_bihash_kv_48_8_t * v)
|
|
{
|
|
#ifdef clib_crc32c_uses_intrinsics
|
|
return clib_crc32c ((u8 *) v->key, 48);
|
|
#else
|
|
u64 tmp = v->key[0] ^ v->key[1] ^ v->key[2] ^ v->key[3] ^ v->key[4]
|
|
^ v->key[5];
|
|
return clib_xxhash (tmp);
|
|
#endif
|
|
}
|
|
|
|
static inline u8 *
|
|
format_bihash_kvp_48_8 (u8 * s, va_list * args)
|
|
{
|
|
clib_bihash_kv_48_8_t *v = va_arg (*args, clib_bihash_kv_48_8_t *);
|
|
|
|
s = format (s, "key %llu %llu %llu %llu %llu %llu value %llu", v->key[0],
|
|
v->key[1], v->key[2], v->key[3], v->key[4], v->key[5],
|
|
v->value);
|
|
return s;
|
|
}
|
|
|
|
static inline int
|
|
clib_bihash_key_compare_48_8 (u64 * a, u64 * b)
|
|
{
|
|
#if defined (CLIB_HAVE_VEC512)
|
|
u64x8 v = u64x8_load_unaligned (a) ^ u64x8_load_unaligned (b);
|
|
return (u64x8_is_zero_mask (v) & 0x3f) == 0;
|
|
#elif defined (CLIB_HAVE_VEC256)
|
|
u64x4 v = { 0 };
|
|
v = u64x4_insert_lo (v, u64x2_load_unaligned (a + 4) ^
|
|
u64x2_load_unaligned (b + 4));
|
|
v |= u64x4_load_unaligned (a) ^ u64x4_load_unaligned (b);
|
|
return u64x4_is_all_zero (v);
|
|
#elif defined(CLIB_HAVE_VEC128) && defined(CLIB_HAVE_VEC128_UNALIGNED_LOAD_STORE)
|
|
u64x2 v;
|
|
v = u64x2_load_unaligned (a) ^ u64x2_load_unaligned (b);
|
|
v |= u64x2_load_unaligned (a + 2) ^ u64x2_load_unaligned (b + 2);
|
|
v |= u64x2_load_unaligned (a + 4) ^ u64x2_load_unaligned (b + 4);
|
|
return u64x2_is_all_zero (v);
|
|
#else
|
|
return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2]) | (a[3] ^ b[3])
|
|
| (a[4] ^ b[4]) | (a[5] ^ b[5])) == 0;
|
|
#endif
|
|
}
|
|
|
|
#undef __included_bihash_template_h__
|
|
#include <vppinfra/bihash_template.h>
|
|
|
|
#endif /* __included_bihash_48_8_h__ */
|
|
|
|
/*
|
|
* fd.io coding-style-patch-verification: ON
|
|
*
|
|
* Local Variables:
|
|
* eval: (c-set-style "gnu")
|
|
* End:
|
|
*/
|