2 Copyright (c) 2003-2017, Troy D. Hanson http://troydhanson.github.com/uthash/
5 Redistribution and use in source and binary forms, with or without
6 modification, are permitted provided that the following conditions are met:
8 * Redistributions of source code must retain the above copyright
9 notice, this list of conditions and the following disclaimer.
11 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
12 IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
13 TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
14 PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER
15 OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
16 EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
17 PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
18 PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
19 LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
20 NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
21 SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 #define UTHASH_VERSION 2.0.2
29 #include <string.h> /* memcmp,strlen */
30 #include <stddef.h> /* ptrdiff_t */
31 #include <stdlib.h> /* exit() */
33 /* These macros use decltype or the earlier __typeof GNU extension.
34 As decltype is only available in newer compilers (VS2010 or gcc 4.3+
35 when compiling c++ source) this code uses whatever method is needed
36 or, for VS2008 where neither is available, uses casting workarounds. */
37 #if defined(_MSC_VER) /* MS compiler */
38 #if _MSC_VER >= 1600 && defined(__cplusplus) /* VS2010 or newer in C++ mode */
39 #define DECLTYPE(x) (decltype(x))
40 #else /* VS2008 or older (or VS2010 in C mode) */
44 #elif defined(__BORLANDC__) || defined(__LCC__) || defined(__WATCOMC__)
47 #else /* GNU, Sun and other compilers */
48 #define DECLTYPE(x) (__typeof(x))
52 #define DECLTYPE_ASSIGN(dst,src) \
54 char **_da_dst = (char**)(&(dst)); \
55 *_da_dst = (char*)(src); \
58 #define DECLTYPE_ASSIGN(dst,src) \
60 (dst) = DECLTYPE(dst)(src); \
64 /* a number of the hash function use uint32_t which isn't defined on Pre VS2010 */
66 #if defined(_MSC_VER) && _MSC_VER >= 1600
68 #elif defined(__WATCOMC__) || defined(__MINGW32__) || defined(__CYGWIN__)
71 typedef unsigned int uint32_t;
72 typedef unsigned char uint8_t;
74 #elif defined(__GNUC__) && !defined(__VXWORKS__)
77 typedef unsigned int uint32_t;
78 typedef unsigned char uint8_t;
82 #define uthash_fatal(msg) exit(-1) /* fatal error (out of memory,etc) */
85 #define uthash_malloc(sz) malloc(sz) /* malloc fcn */
88 #define uthash_free(ptr,sz) free(ptr) /* free fcn */
91 #define uthash_strlen(s) strlen(s)
94 #define uthash_memcmp(a,b,n) memcmp(a,b,n)
97 #ifndef uthash_noexpand_fyi
98 #define uthash_noexpand_fyi(tbl) /* can be defined to log noexpand */
100 #ifndef uthash_expand_fyi
101 #define uthash_expand_fyi(tbl) /* can be defined to log expands */
104 /* initial number of buckets */
105 #define HASH_INITIAL_NUM_BUCKETS 32U /* initial number of buckets */
106 #define HASH_INITIAL_NUM_BUCKETS_LOG2 5U /* lg2 of initial number of buckets */
107 #define HASH_BKT_CAPACITY_THRESH 10U /* expand when bucket count reaches */
109 /* calculate the element whose hash handle address is hhp */
110 #define ELMT_FROM_HH(tbl,hhp) ((void*)(((char*)(hhp)) - ((tbl)->hho)))
111 /* calculate the hash handle from element address elp */
112 #define HH_FROM_ELMT(tbl,elp) ((UT_hash_handle *)(((char*)(elp)) + ((tbl)->hho)))
114 #define HASH_VALUE(keyptr,keylen,hashv) \
116 HASH_FCN(keyptr, keylen, hashv); \
119 #define HASH_FIND_BYHASHVALUE(hh,head,keyptr,keylen,hashval,out) \
124 HASH_TO_BKT(hashval, (head)->hh.tbl->num_buckets, _hf_bkt); \
125 if (HASH_BLOOM_TEST((head)->hh.tbl, hashval) != 0) { \
126 HASH_FIND_IN_BKT((head)->hh.tbl, hh, (head)->hh.tbl->buckets[ _hf_bkt ], keyptr, keylen, hashval, out); \
131 #define HASH_FIND(hh,head,keyptr,keylen,out) \
133 unsigned _hf_hashv; \
134 HASH_VALUE(keyptr, keylen, _hf_hashv); \
135 HASH_FIND_BYHASHVALUE(hh, head, keyptr, keylen, _hf_hashv, out); \
139 #define HASH_BLOOM_BITLEN (1UL << HASH_BLOOM)
140 #define HASH_BLOOM_BYTELEN (HASH_BLOOM_BITLEN/8UL) + (((HASH_BLOOM_BITLEN%8UL)!=0UL) ? 1UL : 0UL)
141 #define HASH_BLOOM_MAKE(tbl) \
143 (tbl)->bloom_nbits = HASH_BLOOM; \
144 (tbl)->bloom_bv = (uint8_t*)uthash_malloc(HASH_BLOOM_BYTELEN); \
145 if (!((tbl)->bloom_bv)) { uthash_fatal( "out of memory"); } \
146 memset((tbl)->bloom_bv, 0, HASH_BLOOM_BYTELEN); \
147 (tbl)->bloom_sig = HASH_BLOOM_SIGNATURE; \
150 #define HASH_BLOOM_FREE(tbl) \
152 uthash_free((tbl)->bloom_bv, HASH_BLOOM_BYTELEN); \
155 #define HASH_BLOOM_BITSET(bv,idx) (bv[(idx)/8U] |= (1U << ((idx)%8U)))
156 #define HASH_BLOOM_BITTEST(bv,idx) (bv[(idx)/8U] & (1U << ((idx)%8U)))
158 #define HASH_BLOOM_ADD(tbl,hashv) \
159 HASH_BLOOM_BITSET((tbl)->bloom_bv, (hashv & (uint32_t)((1ULL << (tbl)->bloom_nbits) - 1U)))
161 #define HASH_BLOOM_TEST(tbl,hashv) \
162 HASH_BLOOM_BITTEST((tbl)->bloom_bv, (hashv & (uint32_t)((1ULL << (tbl)->bloom_nbits) - 1U)))
165 #define HASH_BLOOM_MAKE(tbl)
166 #define HASH_BLOOM_FREE(tbl)
167 #define HASH_BLOOM_ADD(tbl,hashv)
168 #define HASH_BLOOM_TEST(tbl,hashv) (1)
169 #define HASH_BLOOM_BYTELEN 0U
172 #define HASH_MAKE_TABLE(hh,head) \
174 (head)->hh.tbl = (UT_hash_table*)uthash_malloc( \
175 sizeof(UT_hash_table)); \
176 if (!((head)->hh.tbl)) { uthash_fatal( "out of memory"); } \
177 memset((head)->hh.tbl, 0, sizeof(UT_hash_table)); \
178 (head)->hh.tbl->tail = &((head)->hh); \
179 (head)->hh.tbl->num_buckets = HASH_INITIAL_NUM_BUCKETS; \
180 (head)->hh.tbl->log2_num_buckets = HASH_INITIAL_NUM_BUCKETS_LOG2; \
181 (head)->hh.tbl->hho = (char*)(&(head)->hh) - (char*)(head); \
182 (head)->hh.tbl->buckets = (UT_hash_bucket*)uthash_malloc( \
183 HASH_INITIAL_NUM_BUCKETS*sizeof(struct UT_hash_bucket)); \
184 if (! (head)->hh.tbl->buckets) { uthash_fatal( "out of memory"); } \
185 memset((head)->hh.tbl->buckets, 0, \
186 HASH_INITIAL_NUM_BUCKETS*sizeof(struct UT_hash_bucket)); \
187 HASH_BLOOM_MAKE((head)->hh.tbl); \
188 (head)->hh.tbl->signature = HASH_SIGNATURE; \
191 #define HASH_REPLACE_BYHASHVALUE_INORDER(hh,head,fieldname,keylen_in,hashval,add,replaced,cmpfcn) \
194 HASH_FIND_BYHASHVALUE(hh, head, &((add)->fieldname), keylen_in, hashval, replaced); \
196 HASH_DELETE(hh, head, replaced); \
198 HASH_ADD_KEYPTR_BYHASHVALUE_INORDER(hh, head, &((add)->fieldname), keylen_in, hashval, add, cmpfcn); \
201 #define HASH_REPLACE_BYHASHVALUE(hh,head,fieldname,keylen_in,hashval,add,replaced) \
204 HASH_FIND_BYHASHVALUE(hh, head, &((add)->fieldname), keylen_in, hashval, replaced); \
206 HASH_DELETE(hh, head, replaced); \
208 HASH_ADD_KEYPTR_BYHASHVALUE(hh, head, &((add)->fieldname), keylen_in, hashval, add); \
211 #define HASH_REPLACE(hh,head,fieldname,keylen_in,add,replaced) \
213 unsigned _hr_hashv; \
214 HASH_VALUE(&((add)->fieldname), keylen_in, _hr_hashv); \
215 HASH_REPLACE_BYHASHVALUE(hh, head, fieldname, keylen_in, _hr_hashv, add, replaced); \
218 #define HASH_REPLACE_INORDER(hh,head,fieldname,keylen_in,add,replaced,cmpfcn) \
220 unsigned _hr_hashv; \
221 HASH_VALUE(&((add)->fieldname), keylen_in, _hr_hashv); \
222 HASH_REPLACE_BYHASHVALUE_INORDER(hh, head, fieldname, keylen_in, _hr_hashv, add, replaced, cmpfcn); \
225 #define HASH_APPEND_LIST(hh, head, add) \
227 (add)->hh.next = NULL; \
228 (add)->hh.prev = ELMT_FROM_HH((head)->hh.tbl, (head)->hh.tbl->tail); \
229 (head)->hh.tbl->tail->next = (add); \
230 (head)->hh.tbl->tail = &((add)->hh); \
233 #define HASH_ADD_KEYPTR_BYHASHVALUE_INORDER(hh,head,keyptr,keylen_in,hashval,add,cmpfcn) \
236 (add)->hh.hashv = (hashval); \
237 (add)->hh.key = (char*) (keyptr); \
238 (add)->hh.keylen = (unsigned) (keylen_in); \
240 (add)->hh.next = NULL; \
241 (add)->hh.prev = NULL; \
243 HASH_MAKE_TABLE(hh, head); \
245 void *_hs_iter = (head); \
246 (add)->hh.tbl = (head)->hh.tbl; \
248 if (cmpfcn(DECLTYPE(head)(_hs_iter), add) > 0) \
250 } while ((_hs_iter = HH_FROM_ELMT((head)->hh.tbl, _hs_iter)->next)); \
252 (add)->hh.next = _hs_iter; \
253 if (((add)->hh.prev = HH_FROM_ELMT((head)->hh.tbl, _hs_iter)->prev)) { \
254 HH_FROM_ELMT((head)->hh.tbl, (add)->hh.prev)->next = (add); \
258 HH_FROM_ELMT((head)->hh.tbl, _hs_iter)->prev = (add); \
260 HASH_APPEND_LIST(hh, head, add); \
263 (head)->hh.tbl->num_items++; \
264 HASH_TO_BKT(hashval, (head)->hh.tbl->num_buckets, _ha_bkt); \
265 HASH_ADD_TO_BKT((head)->hh.tbl->buckets[_ha_bkt], &(add)->hh); \
266 HASH_BLOOM_ADD((head)->hh.tbl, hashval); \
267 HASH_EMIT_KEY(hh, head, keyptr, keylen_in); \
268 HASH_FSCK(hh, head); \
271 #define HASH_ADD_KEYPTR_INORDER(hh,head,keyptr,keylen_in,add,cmpfcn) \
273 unsigned _hs_hashv; \
274 HASH_VALUE(keyptr, keylen_in, _hs_hashv); \
275 HASH_ADD_KEYPTR_BYHASHVALUE_INORDER(hh, head, keyptr, keylen_in, _hs_hashv, add, cmpfcn); \
278 #define HASH_ADD_BYHASHVALUE_INORDER(hh,head,fieldname,keylen_in,hashval,add,cmpfcn) \
279 HASH_ADD_KEYPTR_BYHASHVALUE_INORDER(hh, head, &((add)->fieldname), keylen_in, hashval, add, cmpfcn)
281 #define HASH_ADD_INORDER(hh,head,fieldname,keylen_in,add,cmpfcn) \
282 HASH_ADD_KEYPTR_INORDER(hh, head, &((add)->fieldname), keylen_in, add, cmpfcn)
284 #define HASH_ADD_KEYPTR_BYHASHVALUE(hh,head,keyptr,keylen_in,hashval,add) \
287 (add)->hh.hashv = (hashval); \
288 (add)->hh.key = (char*) (keyptr); \
289 (add)->hh.keylen = (unsigned) (keylen_in); \
291 (add)->hh.next = NULL; \
292 (add)->hh.prev = NULL; \
294 HASH_MAKE_TABLE(hh, head); \
296 (add)->hh.tbl = (head)->hh.tbl; \
297 HASH_APPEND_LIST(hh, head, add); \
299 (head)->hh.tbl->num_items++; \
300 HASH_TO_BKT(hashval, (head)->hh.tbl->num_buckets, _ha_bkt); \
301 HASH_ADD_TO_BKT((head)->hh.tbl->buckets[_ha_bkt], &(add)->hh); \
302 HASH_BLOOM_ADD((head)->hh.tbl, hashval); \
303 HASH_EMIT_KEY(hh, head, keyptr, keylen_in); \
304 HASH_FSCK(hh, head); \
307 #define HASH_ADD_KEYPTR(hh,head,keyptr,keylen_in,add) \
309 unsigned _ha_hashv; \
310 HASH_VALUE(keyptr, keylen_in, _ha_hashv); \
311 HASH_ADD_KEYPTR_BYHASHVALUE(hh, head, keyptr, keylen_in, _ha_hashv, add); \
314 #define HASH_ADD_BYHASHVALUE(hh,head,fieldname,keylen_in,hashval,add) \
315 HASH_ADD_KEYPTR_BYHASHVALUE(hh, head, &((add)->fieldname), keylen_in, hashval, add)
317 #define HASH_ADD(hh,head,fieldname,keylen_in,add) \
318 HASH_ADD_KEYPTR(hh, head, &((add)->fieldname), keylen_in, add)
320 #define HASH_TO_BKT(hashv,num_bkts,bkt) \
322 bkt = ((hashv) & ((num_bkts) - 1U)); \
325 /* delete "delptr" from the hash table.
326 * "the usual" patch-up process for the app-order doubly-linked-list.
327 * The use of _hd_hh_del below deserves special explanation.
328 * These used to be expressed using (delptr) but that led to a bug
329 * if someone used the same symbol for the head and deletee, like
330 * HASH_DELETE(hh,users,users);
331 * We want that to work, but by changing the head (users) below
332 * we were forfeiting our ability to further refer to the deletee (users)
333 * in the patch-up process. Solution: use scratch space to
334 * copy the deletee pointer, then the latter references are via that
335 * scratch pointer rather than through the repointed (users) symbol.
337 #define HASH_DELETE(hh,head,delptr) \
339 struct UT_hash_handle *_hd_hh_del; \
340 if ( ((delptr)->hh.prev == NULL) && ((delptr)->hh.next == NULL) ) { \
341 uthash_free((head)->hh.tbl->buckets, \
342 (head)->hh.tbl->num_buckets*sizeof(struct UT_hash_bucket) ); \
343 HASH_BLOOM_FREE((head)->hh.tbl); \
344 uthash_free((head)->hh.tbl, sizeof(UT_hash_table)); \
348 _hd_hh_del = &((delptr)->hh); \
349 if ((delptr) == ELMT_FROM_HH((head)->hh.tbl,(head)->hh.tbl->tail)) { \
350 (head)->hh.tbl->tail = \
351 (UT_hash_handle*)((ptrdiff_t)((delptr)->hh.prev) + \
352 (head)->hh.tbl->hho); \
354 if ((delptr)->hh.prev != NULL) { \
355 ((UT_hash_handle*)((ptrdiff_t)((delptr)->hh.prev) + \
356 (head)->hh.tbl->hho))->next = (delptr)->hh.next; \
358 DECLTYPE_ASSIGN(head,(delptr)->hh.next); \
360 if (_hd_hh_del->next != NULL) { \
361 ((UT_hash_handle*)((ptrdiff_t)_hd_hh_del->next + \
362 (head)->hh.tbl->hho))->prev = \
365 HASH_TO_BKT( _hd_hh_del->hashv, (head)->hh.tbl->num_buckets, _hd_bkt); \
366 HASH_DEL_IN_BKT(hh,(head)->hh.tbl->buckets[_hd_bkt], _hd_hh_del); \
367 (head)->hh.tbl->num_items--; \
369 HASH_FSCK(hh,head); \
373 /* convenience forms of HASH_FIND/HASH_ADD/HASH_DEL */
374 #define HASH_FIND_STR(head,findstr,out) \
375 HASH_FIND(hh,head,findstr,(unsigned)uthash_strlen(findstr),out)
376 #define HASH_ADD_STR(head,strfield,add) \
377 HASH_ADD(hh,head,strfield[0],(unsigned)uthash_strlen(add->strfield),add)
378 #define HASH_REPLACE_STR(head,strfield,add,replaced) \
379 HASH_REPLACE(hh,head,strfield[0],(unsigned)uthash_strlen(add->strfield),add,replaced)
380 #define HASH_FIND_INT(head,findint,out) \
381 HASH_FIND(hh,head,findint,sizeof(int),out)
382 #define HASH_ADD_INT(head,intfield,add) \
383 HASH_ADD(hh,head,intfield,sizeof(int),add)
384 #define HASH_REPLACE_INT(head,intfield,add,replaced) \
385 HASH_REPLACE(hh,head,intfield,sizeof(int),add,replaced)
386 #define HASH_FIND_PTR(head,findptr,out) \
387 HASH_FIND(hh,head,findptr,sizeof(void *),out)
388 #define HASH_ADD_PTR(head,ptrfield,add) \
389 HASH_ADD(hh,head,ptrfield,sizeof(void *),add)
390 #define HASH_REPLACE_PTR(head,ptrfield,add,replaced) \
391 HASH_REPLACE(hh,head,ptrfield,sizeof(void *),add,replaced)
392 #define HASH_DEL(head,delptr) \
393 HASH_DELETE(hh,head,delptr)
395 /* HASH_FSCK checks hash integrity on every add/delete when HASH_DEBUG is defined.
396 * This is for uthash developer only; it compiles away if HASH_DEBUG isn't defined.
399 #define HASH_OOPS(...) do { fprintf(stderr,__VA_ARGS__); exit(-1); } while (0)
400 #define HASH_FSCK(hh,head) \
402 struct UT_hash_handle *_thh; \
408 for( _bkt_i = 0; _bkt_i < (head)->hh.tbl->num_buckets; _bkt_i++) { \
409 unsigned _bkt_count = 0; \
410 _thh = (head)->hh.tbl->buckets[_bkt_i].hh_head; \
413 if (_prev != (char*)(_thh->hh_prev)) { \
414 HASH_OOPS("invalid hh_prev %p, actual %p\n", \
415 _thh->hh_prev, _prev ); \
418 _prev = (char*)(_thh); \
419 _thh = _thh->hh_next; \
421 _count += _bkt_count; \
422 if ((head)->hh.tbl->buckets[_bkt_i].count != _bkt_count) { \
423 HASH_OOPS("invalid bucket count %u, actual %u\n", \
424 (head)->hh.tbl->buckets[_bkt_i].count, _bkt_count); \
427 if (_count != (head)->hh.tbl->num_items) { \
428 HASH_OOPS("invalid hh item count %u, actual %u\n", \
429 (head)->hh.tbl->num_items, _count ); \
431 /* traverse hh in app order; check next/prev integrity, count */ \
434 _thh = &(head)->hh; \
437 if (_prev !=(char*)(_thh->prev)) { \
438 HASH_OOPS("invalid prev %p, actual %p\n", \
439 _thh->prev, _prev ); \
441 _prev = (char*)ELMT_FROM_HH((head)->hh.tbl, _thh); \
442 _thh = ( _thh->next ? (UT_hash_handle*)((char*)(_thh->next) + \
443 (head)->hh.tbl->hho) : NULL ); \
445 if (_count != (head)->hh.tbl->num_items) { \
446 HASH_OOPS("invalid app item count %u, actual %u\n", \
447 (head)->hh.tbl->num_items, _count ); \
452 #define HASH_FSCK(hh,head)
455 /* When compiled with -DHASH_EMIT_KEYS, length-prefixed keys are emitted to
456 * the descriptor to which this macro is defined for tuning the hash function.
457 * The app can #include <unistd.h> to get the prototype for write(2). */
458 #ifdef HASH_EMIT_KEYS
459 #define HASH_EMIT_KEY(hh,head,keyptr,fieldlen) \
461 unsigned _klen = fieldlen; \
462 write(HASH_EMIT_KEYS, &_klen, sizeof(_klen)); \
463 write(HASH_EMIT_KEYS, keyptr, (unsigned long)fieldlen); \
466 #define HASH_EMIT_KEY(hh,head,keyptr,fieldlen)
469 /* default to Jenkin's hash unless overridden e.g. DHASH_FUNCTION=HASH_SAX */
471 #define HASH_FCN HASH_FUNCTION
473 #define HASH_FCN HASH_JEN
476 /* The Bernstein hash function, used in Perl prior to v5.6. Note (x<<5+x)=x*33. */
477 #define HASH_BER(key,keylen,hashv) \
479 unsigned _hb_keylen=(unsigned)keylen; \
480 const unsigned char *_hb_key=(const unsigned char*)(key); \
482 while (_hb_keylen-- != 0U) { \
483 (hashv) = (((hashv) << 5) + (hashv)) + *_hb_key++; \
488 /* SAX/FNV/OAT/JEN hash functions are macro variants of those listed at
489 * http://eternallyconfuzzled.com/tuts/algorithms/jsw_tut_hashing.aspx */
490 #define HASH_SAX(key,keylen,hashv) \
493 const unsigned char *_hs_key=(const unsigned char*)(key); \
495 for(_sx_i=0; _sx_i < keylen; _sx_i++) { \
496 hashv ^= (hashv << 5) + (hashv >> 2) + _hs_key[_sx_i]; \
499 /* FNV-1a variation */
500 #define HASH_FNV(key,keylen,hashv) \
503 const unsigned char *_hf_key=(const unsigned char*)(key); \
504 hashv = 2166136261U; \
505 for(_fn_i=0; _fn_i < keylen; _fn_i++) { \
506 hashv = hashv ^ _hf_key[_fn_i]; \
507 hashv = hashv * 16777619U; \
511 #define HASH_OAT(key,keylen,hashv) \
514 const unsigned char *_ho_key=(const unsigned char*)(key); \
516 for(_ho_i=0; _ho_i < keylen; _ho_i++) { \
517 hashv += _ho_key[_ho_i]; \
518 hashv += (hashv << 10); \
519 hashv ^= (hashv >> 6); \
521 hashv += (hashv << 3); \
522 hashv ^= (hashv >> 11); \
523 hashv += (hashv << 15); \
526 #define HASH_JEN_MIX(a,b,c) \
528 a -= b; a -= c; a ^= ( c >> 13 ); \
529 b -= c; b -= a; b ^= ( a << 8 ); \
530 c -= a; c -= b; c ^= ( b >> 13 ); \
531 a -= b; a -= c; a ^= ( c >> 12 ); \
532 b -= c; b -= a; b ^= ( a << 16 ); \
533 c -= a; c -= b; c ^= ( b >> 5 ); \
534 a -= b; a -= c; a ^= ( c >> 3 ); \
535 b -= c; b -= a; b ^= ( a << 10 ); \
536 c -= a; c -= b; c ^= ( b >> 15 ); \
539 #define HASH_JEN(key,keylen,hashv) \
541 unsigned _hj_i,_hj_j,_hj_k; \
542 unsigned const char *_hj_key=(unsigned const char*)(key); \
543 hashv = 0xfeedbeefu; \
544 _hj_i = _hj_j = 0x9e3779b9u; \
545 _hj_k = (unsigned)(keylen); \
546 while (_hj_k >= 12U) { \
547 _hj_i += (_hj_key[0] + ( (unsigned)_hj_key[1] << 8 ) \
548 + ( (unsigned)_hj_key[2] << 16 ) \
549 + ( (unsigned)_hj_key[3] << 24 ) ); \
550 _hj_j += (_hj_key[4] + ( (unsigned)_hj_key[5] << 8 ) \
551 + ( (unsigned)_hj_key[6] << 16 ) \
552 + ( (unsigned)_hj_key[7] << 24 ) ); \
553 hashv += (_hj_key[8] + ( (unsigned)_hj_key[9] << 8 ) \
554 + ( (unsigned)_hj_key[10] << 16 ) \
555 + ( (unsigned)_hj_key[11] << 24 ) ); \
557 HASH_JEN_MIX(_hj_i, _hj_j, hashv); \
562 hashv += (unsigned)(keylen); \
564 case 11: hashv += ( (unsigned)_hj_key[10] << 24 ); /* FALLTHROUGH */ \
565 case 10: hashv += ( (unsigned)_hj_key[9] << 16 ); /* FALLTHROUGH */ \
566 case 9: hashv += ( (unsigned)_hj_key[8] << 8 ); /* FALLTHROUGH */ \
567 case 8: _hj_j += ( (unsigned)_hj_key[7] << 24 ); /* FALLTHROUGH */ \
568 case 7: _hj_j += ( (unsigned)_hj_key[6] << 16 ); /* FALLTHROUGH */ \
569 case 6: _hj_j += ( (unsigned)_hj_key[5] << 8 ); /* FALLTHROUGH */ \
570 case 5: _hj_j += _hj_key[4]; /* FALLTHROUGH */ \
571 case 4: _hj_i += ( (unsigned)_hj_key[3] << 24 ); /* FALLTHROUGH */ \
572 case 3: _hj_i += ( (unsigned)_hj_key[2] << 16 ); /* FALLTHROUGH */ \
573 case 2: _hj_i += ( (unsigned)_hj_key[1] << 8 ); /* FALLTHROUGH */ \
574 case 1: _hj_i += _hj_key[0]; \
576 HASH_JEN_MIX(_hj_i, _hj_j, hashv); \
579 /* The Paul Hsieh hash function */
581 #if (defined(__GNUC__) && defined(__i386__)) || defined(__WATCOMC__) \
582 || defined(_MSC_VER) || defined (__BORLANDC__) || defined (__TURBOC__)
583 #define get16bits(d) (*((const uint16_t *) (d)))
586 #if !defined (get16bits)
587 #define get16bits(d) ((((uint32_t)(((const uint8_t *)(d))[1])) << 8) \
588 +(uint32_t)(((const uint8_t *)(d))[0]) )
590 #define HASH_SFH(key,keylen,hashv) \
592 unsigned const char *_sfh_key=(unsigned const char*)(key); \
593 uint32_t _sfh_tmp, _sfh_len = (uint32_t)keylen; \
595 unsigned _sfh_rem = _sfh_len & 3U; \
597 hashv = 0xcafebabeu; \
600 for (;_sfh_len > 0U; _sfh_len--) { \
601 hashv += get16bits (_sfh_key); \
602 _sfh_tmp = ((uint32_t)(get16bits (_sfh_key+2)) << 11) ^ hashv; \
603 hashv = (hashv << 16) ^ _sfh_tmp; \
604 _sfh_key += 2U*sizeof (uint16_t); \
605 hashv += hashv >> 11; \
608 /* Handle end cases */ \
609 switch (_sfh_rem) { \
610 case 3: hashv += get16bits (_sfh_key); \
611 hashv ^= hashv << 16; \
612 hashv ^= (uint32_t)(_sfh_key[sizeof (uint16_t)]) << 18; \
613 hashv += hashv >> 11; \
615 case 2: hashv += get16bits (_sfh_key); \
616 hashv ^= hashv << 11; \
617 hashv += hashv >> 17; \
619 case 1: hashv += *_sfh_key; \
620 hashv ^= hashv << 10; \
621 hashv += hashv >> 1; \
624 /* Force "avalanching" of final 127 bits */ \
625 hashv ^= hashv << 3; \
626 hashv += hashv >> 5; \
627 hashv ^= hashv << 4; \
628 hashv += hashv >> 17; \
629 hashv ^= hashv << 25; \
630 hashv += hashv >> 6; \
633 #ifdef HASH_USING_NO_STRICT_ALIASING
634 /* The MurmurHash exploits some CPU's (x86,x86_64) tolerance for unaligned reads.
635 * For other types of CPU's (e.g. Sparc) an unaligned read causes a bus error.
636 * MurmurHash uses the faster approach only on CPU's where we know it's safe.
638 * Note the preprocessor built-in defines can be emitted using:
640 * gcc -m64 -dM -E - < /dev/null (on gcc)
641 * cc -## a.c (where a.c is a simple test file) (Sun Studio)
643 #if (defined(__i386__) || defined(__x86_64__) || defined(_M_IX86))
644 #define MUR_GETBLOCK(p,i) p[i]
645 #else /* non intel */
646 #define MUR_PLUS0_ALIGNED(p) (((unsigned long)p & 3UL) == 0UL)
647 #define MUR_PLUS1_ALIGNED(p) (((unsigned long)p & 3UL) == 1UL)
648 #define MUR_PLUS2_ALIGNED(p) (((unsigned long)p & 3UL) == 2UL)
649 #define MUR_PLUS3_ALIGNED(p) (((unsigned long)p & 3UL) == 3UL)
650 #define WP(p) ((uint32_t*)((unsigned long)(p) & ~3UL))
651 #if (defined(__BIG_ENDIAN__) || defined(SPARC) || defined(__ppc__) || defined(__ppc64__))
652 #define MUR_THREE_ONE(p) ((((*WP(p))&0x00ffffff) << 8) | (((*(WP(p)+1))&0xff000000) >> 24))
653 #define MUR_TWO_TWO(p) ((((*WP(p))&0x0000ffff) <<16) | (((*(WP(p)+1))&0xffff0000) >> 16))
654 #define MUR_ONE_THREE(p) ((((*WP(p))&0x000000ff) <<24) | (((*(WP(p)+1))&0xffffff00) >> 8))
655 #else /* assume little endian non-intel */
656 #define MUR_THREE_ONE(p) ((((*WP(p))&0xffffff00) >> 8) | (((*(WP(p)+1))&0x000000ff) << 24))
657 #define MUR_TWO_TWO(p) ((((*WP(p))&0xffff0000) >>16) | (((*(WP(p)+1))&0x0000ffff) << 16))
658 #define MUR_ONE_THREE(p) ((((*WP(p))&0xff000000) >>24) | (((*(WP(p)+1))&0x00ffffff) << 8))
660 #define MUR_GETBLOCK(p,i) (MUR_PLUS0_ALIGNED(p) ? ((p)[i]) : \
661 (MUR_PLUS1_ALIGNED(p) ? MUR_THREE_ONE(p) : \
662 (MUR_PLUS2_ALIGNED(p) ? MUR_TWO_TWO(p) : \
665 #define MUR_ROTL32(x,r) (((x) << (r)) | ((x) >> (32 - (r))))
666 #define MUR_FMIX(_h) \
675 #define HASH_MUR(key,keylen,hashv) \
677 const uint8_t *_mur_data = (const uint8_t*)(key); \
678 const int _mur_nblocks = (int)(keylen) / 4; \
679 uint32_t _mur_h1 = 0xf88D5353u; \
680 uint32_t _mur_c1 = 0xcc9e2d51u; \
681 uint32_t _mur_c2 = 0x1b873593u; \
682 uint32_t _mur_k1 = 0; \
683 const uint8_t *_mur_tail; \
684 const uint32_t *_mur_blocks = (const uint32_t*)(_mur_data+(_mur_nblocks*4)); \
686 for(_mur_i = -_mur_nblocks; _mur_i!=0; _mur_i++) { \
687 _mur_k1 = MUR_GETBLOCK(_mur_blocks,_mur_i); \
688 _mur_k1 *= _mur_c1; \
689 _mur_k1 = MUR_ROTL32(_mur_k1,15); \
690 _mur_k1 *= _mur_c2; \
692 _mur_h1 ^= _mur_k1; \
693 _mur_h1 = MUR_ROTL32(_mur_h1,13); \
694 _mur_h1 = (_mur_h1*5U) + 0xe6546b64u; \
696 _mur_tail = (const uint8_t*)(_mur_data + (_mur_nblocks*4)); \
698 switch((keylen) & 3U) { \
699 case 3: _mur_k1 ^= (uint32_t)_mur_tail[2] << 16; /* FALLTHROUGH */ \
700 case 2: _mur_k1 ^= (uint32_t)_mur_tail[1] << 8; /* FALLTHROUGH */ \
701 case 1: _mur_k1 ^= (uint32_t)_mur_tail[0]; \
702 _mur_k1 *= _mur_c1; \
703 _mur_k1 = MUR_ROTL32(_mur_k1,15); \
704 _mur_k1 *= _mur_c2; \
705 _mur_h1 ^= _mur_k1; \
707 _mur_h1 ^= (uint32_t)(keylen); \
711 #endif /* HASH_USING_NO_STRICT_ALIASING */
713 /* iterate over items in a known bucket to find desired item */
714 #define HASH_FIND_IN_BKT(tbl,hh,head,keyptr,keylen_in,hashval,out) \
716 if ((head).hh_head != NULL) { \
717 DECLTYPE_ASSIGN(out, ELMT_FROM_HH(tbl, (head).hh_head)); \
721 while ((out) != NULL) { \
722 if ((out)->hh.hashv == (hashval) && (out)->hh.keylen == (keylen_in)) { \
723 if (uthash_memcmp((out)->hh.key, keyptr, keylen_in) == 0) { \
727 if ((out)->hh.hh_next != NULL) { \
728 DECLTYPE_ASSIGN(out, ELMT_FROM_HH(tbl, (out)->hh.hh_next)); \
735 /* add an item to a bucket */
736 #define HASH_ADD_TO_BKT(head,addhh) \
739 (addhh)->hh_next = head.hh_head; \
740 (addhh)->hh_prev = NULL; \
741 if (head.hh_head != NULL) { (head).hh_head->hh_prev = (addhh); } \
742 (head).hh_head=addhh; \
743 if ((head.count >= ((head.expand_mult+1U) * HASH_BKT_CAPACITY_THRESH)) \
744 && ((addhh)->tbl->noexpand != 1U)) { \
745 HASH_EXPAND_BUCKETS((addhh)->tbl); \
749 /* remove an item from a given bucket */
750 #define HASH_DEL_IN_BKT(hh,head,hh_del) \
752 if ((head).hh_head == hh_del) { \
753 (head).hh_head = hh_del->hh_next; \
755 if (hh_del->hh_prev) { \
756 hh_del->hh_prev->hh_next = hh_del->hh_next; \
758 if (hh_del->hh_next) { \
759 hh_del->hh_next->hh_prev = hh_del->hh_prev; \
762 /* Bucket expansion has the effect of doubling the number of buckets
763 * and redistributing the items into the new buckets. Ideally the
764 * items will distribute more or less evenly into the new buckets
765 * (the extent to which this is true is a measure of the quality of
766 * the hash function as it applies to the key domain).
768 * With the items distributed into more buckets, the chain length
769 * (item count) in each bucket is reduced. Thus by expanding buckets
770 * the hash keeps a bound on the chain length. This bounded chain
771 * length is the essence of how a hash provides constant time lookup.
773 * The calculation of tbl->ideal_chain_maxlen below deserves some
774 * explanation. First, keep in mind that we're calculating the ideal
775 * maximum chain length based on the *new* (doubled) bucket count.
776 * In fractions this is just n/b (n=number of items,b=new num buckets).
777 * Since the ideal chain length is an integer, we want to calculate
778 * ceil(n/b). We don't depend on floating point arithmetic in this
779 * hash, so to calculate ceil(n/b) with integers we could write
781 * ceil(n/b) = (n/b) + ((n%b)?1:0)
783 * and in fact a previous version of this hash did just that.
784 * But now we have improved things a bit by recognizing that b is
785 * always a power of two. We keep its base 2 log handy (call it lb),
786 * so now we can write this with a bit shift and logical AND:
788 * ceil(n/b) = (n>>lb) + ( (n & (b-1)) ? 1:0)
791 #define HASH_EXPAND_BUCKETS(tbl) \
794 unsigned _he_bkt_i; \
795 struct UT_hash_handle *_he_thh, *_he_hh_nxt; \
796 UT_hash_bucket *_he_new_buckets, *_he_newbkt; \
797 _he_new_buckets = (UT_hash_bucket*)uthash_malloc( \
798 2UL * tbl->num_buckets * sizeof(struct UT_hash_bucket)); \
799 if (!_he_new_buckets) { uthash_fatal( "out of memory"); } \
800 memset(_he_new_buckets, 0, \
801 2UL * tbl->num_buckets * sizeof(struct UT_hash_bucket)); \
802 tbl->ideal_chain_maxlen = \
803 (tbl->num_items >> (tbl->log2_num_buckets+1U)) + \
804 (((tbl->num_items & ((tbl->num_buckets*2U)-1U)) != 0U) ? 1U : 0U); \
805 tbl->nonideal_items = 0; \
806 for(_he_bkt_i = 0; _he_bkt_i < tbl->num_buckets; _he_bkt_i++) \
808 _he_thh = tbl->buckets[ _he_bkt_i ].hh_head; \
809 while (_he_thh != NULL) { \
810 _he_hh_nxt = _he_thh->hh_next; \
811 HASH_TO_BKT( _he_thh->hashv, tbl->num_buckets*2U, _he_bkt); \
812 _he_newbkt = &(_he_new_buckets[ _he_bkt ]); \
813 if (++(_he_newbkt->count) > tbl->ideal_chain_maxlen) { \
814 tbl->nonideal_items++; \
815 _he_newbkt->expand_mult = _he_newbkt->count / \
816 tbl->ideal_chain_maxlen; \
818 _he_thh->hh_prev = NULL; \
819 _he_thh->hh_next = _he_newbkt->hh_head; \
820 if (_he_newbkt->hh_head != NULL) { _he_newbkt->hh_head->hh_prev = \
822 _he_newbkt->hh_head = _he_thh; \
823 _he_thh = _he_hh_nxt; \
826 uthash_free( tbl->buckets, tbl->num_buckets*sizeof(struct UT_hash_bucket) ); \
827 tbl->num_buckets *= 2U; \
828 tbl->log2_num_buckets++; \
829 tbl->buckets = _he_new_buckets; \
830 tbl->ineff_expands = (tbl->nonideal_items > (tbl->num_items >> 1)) ? \
831 (tbl->ineff_expands+1U) : 0U; \
832 if (tbl->ineff_expands > 1U) { \
834 uthash_noexpand_fyi(tbl); \
836 uthash_expand_fyi(tbl); \
840 /* This is an adaptation of Simon Tatham's O(n log(n)) mergesort */
841 /* Note that HASH_SORT assumes the hash handle name to be hh.
842 * HASH_SRT was added to allow the hash handle name to be passed in. */
843 #define HASH_SORT(head,cmpfcn) HASH_SRT(hh,head,cmpfcn)
844 #define HASH_SRT(hh,head,cmpfcn) \
847 unsigned _hs_looping,_hs_nmerges,_hs_insize,_hs_psize,_hs_qsize; \
848 struct UT_hash_handle *_hs_p, *_hs_q, *_hs_e, *_hs_list, *_hs_tail; \
849 if (head != NULL) { \
852 _hs_list = &((head)->hh); \
853 while (_hs_looping != 0U) { \
858 while (_hs_p != NULL) { \
862 for ( _hs_i = 0; _hs_i < _hs_insize; _hs_i++ ) { \
864 _hs_q = (UT_hash_handle*)((_hs_q->next != NULL) ? \
865 ((void*)((char*)(_hs_q->next) + \
866 (head)->hh.tbl->hho)) : NULL); \
867 if (! (_hs_q) ) { break; } \
869 _hs_qsize = _hs_insize; \
870 while ((_hs_psize > 0U) || ((_hs_qsize > 0U) && (_hs_q != NULL))) {\
871 if (_hs_psize == 0U) { \
873 _hs_q = (UT_hash_handle*)((_hs_q->next != NULL) ? \
874 ((void*)((char*)(_hs_q->next) + \
875 (head)->hh.tbl->hho)) : NULL); \
877 } else if ( (_hs_qsize == 0U) || (_hs_q == NULL) ) { \
879 if (_hs_p != NULL){ \
880 _hs_p = (UT_hash_handle*)((_hs_p->next != NULL) ? \
881 ((void*)((char*)(_hs_p->next) + \
882 (head)->hh.tbl->hho)) : NULL); \
886 cmpfcn(DECLTYPE(head)(ELMT_FROM_HH((head)->hh.tbl,_hs_p)), \
887 DECLTYPE(head)(ELMT_FROM_HH((head)->hh.tbl,_hs_q))) \
890 if (_hs_p != NULL){ \
891 _hs_p = (UT_hash_handle*)((_hs_p->next != NULL) ? \
892 ((void*)((char*)(_hs_p->next) + \
893 (head)->hh.tbl->hho)) : NULL); \
898 _hs_q = (UT_hash_handle*)((_hs_q->next != NULL) ? \
899 ((void*)((char*)(_hs_q->next) + \
900 (head)->hh.tbl->hho)) : NULL); \
903 if ( _hs_tail != NULL ) { \
904 _hs_tail->next = ((_hs_e != NULL) ? \
905 ELMT_FROM_HH((head)->hh.tbl,_hs_e) : NULL); \
909 if (_hs_e != NULL) { \
910 _hs_e->prev = ((_hs_tail != NULL) ? \
911 ELMT_FROM_HH((head)->hh.tbl,_hs_tail) : NULL); \
917 if (_hs_tail != NULL){ \
918 _hs_tail->next = NULL; \
920 if ( _hs_nmerges <= 1U ) { \
922 (head)->hh.tbl->tail = _hs_tail; \
923 DECLTYPE_ASSIGN(head,ELMT_FROM_HH((head)->hh.tbl, _hs_list)); \
927 HASH_FSCK(hh,head); \
931 /* This function selects items from one hash into another hash.
932 * The end result is that the selected items have dual presence
933 * in both hashes. There is no copy of the items made; rather
934 * they are added into the new hash through a secondary hash
935 * hash handle that must be present in the structure. */
936 #define HASH_SELECT(hh_dst, dst, hh_src, src, cond) \
938 unsigned _src_bkt, _dst_bkt; \
939 void *_last_elt=NULL, *_elt; \
940 UT_hash_handle *_src_hh, *_dst_hh, *_last_elt_hh=NULL; \
941 ptrdiff_t _dst_hho = ((char*)(&(dst)->hh_dst) - (char*)(dst)); \
943 for(_src_bkt=0; _src_bkt < (src)->hh_src.tbl->num_buckets; _src_bkt++) { \
944 for(_src_hh = (src)->hh_src.tbl->buckets[_src_bkt].hh_head; \
946 _src_hh = _src_hh->hh_next) { \
947 _elt = ELMT_FROM_HH((src)->hh_src.tbl, _src_hh); \
949 _dst_hh = (UT_hash_handle*)(((char*)_elt) + _dst_hho); \
950 _dst_hh->key = _src_hh->key; \
951 _dst_hh->keylen = _src_hh->keylen; \
952 _dst_hh->hashv = _src_hh->hashv; \
953 _dst_hh->prev = _last_elt; \
954 _dst_hh->next = NULL; \
955 if (_last_elt_hh != NULL) { _last_elt_hh->next = _elt; } \
957 DECLTYPE_ASSIGN(dst,_elt); \
958 HASH_MAKE_TABLE(hh_dst,dst); \
960 _dst_hh->tbl = (dst)->hh_dst.tbl; \
962 HASH_TO_BKT(_dst_hh->hashv, _dst_hh->tbl->num_buckets, _dst_bkt); \
963 HASH_ADD_TO_BKT(_dst_hh->tbl->buckets[_dst_bkt],_dst_hh); \
964 (dst)->hh_dst.tbl->num_items++; \
966 _last_elt_hh = _dst_hh; \
971 HASH_FSCK(hh_dst,dst); \
974 #define HASH_CLEAR(hh,head) \
976 if (head != NULL) { \
977 uthash_free((head)->hh.tbl->buckets, \
978 (head)->hh.tbl->num_buckets*sizeof(struct UT_hash_bucket)); \
979 HASH_BLOOM_FREE((head)->hh.tbl); \
980 uthash_free((head)->hh.tbl, sizeof(UT_hash_table)); \
985 #define HASH_OVERHEAD(hh,head) \
986 ((head != NULL) ? ( \
987 (size_t)(((head)->hh.tbl->num_items * sizeof(UT_hash_handle)) + \
988 ((head)->hh.tbl->num_buckets * sizeof(UT_hash_bucket)) + \
989 sizeof(UT_hash_table) + \
990 (HASH_BLOOM_BYTELEN))) : 0U)
993 #define HASH_ITER(hh,head,el,tmp) \
994 for(((el)=(head)), ((*(char**)(&(tmp)))=(char*)((head!=NULL)?(head)->hh.next:NULL)); \
995 (el) != NULL; ((el)=(tmp)), ((*(char**)(&(tmp)))=(char*)((tmp!=NULL)?(tmp)->hh.next:NULL)))
997 #define HASH_ITER(hh,head,el,tmp) \
998 for(((el)=(head)), ((tmp)=DECLTYPE(el)((head!=NULL)?(head)->hh.next:NULL)); \
999 (el) != NULL; ((el)=(tmp)), ((tmp)=DECLTYPE(el)((tmp!=NULL)?(tmp)->hh.next:NULL)))
1002 /* obtain a count of items in the hash */
1003 #define HASH_COUNT(head) HASH_CNT(hh,head)
1004 #define HASH_CNT(hh,head) ((head != NULL)?((head)->hh.tbl->num_items):0U)
1006 typedef struct UT_hash_bucket
{
1007 struct UT_hash_handle
*hh_head
;
1010 /* expand_mult is normally set to 0. In this situation, the max chain length
1011 * threshold is enforced at its default value, HASH_BKT_CAPACITY_THRESH. (If
1012 * the bucket's chain exceeds this length, bucket expansion is triggered).
1013 * However, setting expand_mult to a non-zero value delays bucket expansion
1014 * (that would be triggered by additions to this particular bucket)
1015 * until its chain length reaches a *multiple* of HASH_BKT_CAPACITY_THRESH.
1016 * (The multiplier is simply expand_mult+1). The whole idea of this
1017 * multiplier is to reduce bucket expansions, since they are expensive, in
1018 * situations where we know that a particular bucket tends to be overused.
1019 * It is better to let its chain length grow to a longer yet-still-bounded
1020 * value, than to do an O(n) bucket expansion too often.
1022 unsigned expand_mult
;
1026 /* random signature used only to find hash tables in external analysis */
1027 #define HASH_SIGNATURE 0xa0111fe1u
1028 #define HASH_BLOOM_SIGNATURE 0xb12220f2u
1030 typedef struct UT_hash_table
{
1031 UT_hash_bucket
*buckets
;
1032 unsigned num_buckets
, log2_num_buckets
;
1034 struct UT_hash_handle
*tail
; /* tail hh in app order, for fast append */
1035 ptrdiff_t hho
; /* hash handle offset (byte pos of hash handle in element */
1037 /* in an ideal situation (all buckets used equally), no bucket would have
1038 * more than ceil(#items/#buckets) items. that's the ideal chain length. */
1039 unsigned ideal_chain_maxlen
;
1041 /* nonideal_items is the number of items in the hash whose chain position
1042 * exceeds the ideal chain maxlen. these items pay the penalty for an uneven
1043 * hash distribution; reaching them in a chain traversal takes >ideal steps */
1044 unsigned nonideal_items
;
1046 /* ineffective expands occur when a bucket doubling was performed, but
1047 * afterward, more than half the items in the hash had nonideal chain
1048 * positions. If this happens on two consecutive expansions we inhibit any
1049 * further expansion, as it's not helping; this happens when the hash
1050 * function isn't a good fit for the key domain. When expansion is inhibited
1051 * the hash will still work, albeit no longer in constant time. */
1052 unsigned ineff_expands
, noexpand
;
1054 uint32_t signature
; /* used only to find hash tables in external analysis */
1056 uint32_t bloom_sig
; /* used only to test bloom exists in external analysis */
1058 uint8_t bloom_nbits
;
1063 typedef struct UT_hash_handle
{
1064 struct UT_hash_table
*tbl
;
1065 void *prev
; /* prev element in app order */
1066 void *next
; /* next element in app order */
1067 struct UT_hash_handle
*hh_prev
; /* previous hh in bucket order */
1068 struct UT_hash_handle
*hh_next
; /* next hh in bucket order */
1069 void *key
; /* ptr to enclosing struct's key */
1070 unsigned keylen
; /* enclosing struct's key len */
1071 unsigned hashv
; /* result of hash-fcn(key) */
1074 #endif /* UTHASH_H */