X-Git-Url: http://git.efficios.com/?p=lttng-tools.git;a=blobdiff_plain;f=src%2Fcommon%2Fhashtable%2Frculfhash.c;h=9baf4079877393a1289fdcca8a518286e530cddb;hp=06fec785e8ba0e23eee72907e8fb7cce1d2473a1;hb=7567352fb68f5c3f49f549c579f5bd27c883bed2;hpb=cd5376da8a99eaaa06a288c43e18ee852e2d49f2 diff --git a/src/common/hashtable/rculfhash.c b/src/common/hashtable/rculfhash.c index 06fec785e..9baf40798 100644 --- a/src/common/hashtable/rculfhash.c +++ b/src/common/hashtable/rculfhash.c @@ -278,13 +278,7 @@ #include "rculfhash-internal.h" #include "urcu-flavor.h" -/* - * We need to lock pthread exit, which deadlocks __nptl_setxid in the runas - * clone. This work-around will be allowed to be removed when runas.c gets - * changed to do an exec() before issuing seteuid/setegid. See - * http://sourceware.org/bugzilla/show_bug.cgi?id=10184 for details. - */ -pthread_mutex_t lttng_libc_state_lock = PTHREAD_MUTEX_INITIALIZER; +#include /* * Split-counters lazily update the global counter each 1024 @@ -570,6 +564,7 @@ void cds_lfht_resize_lazy_count(struct cds_lfht *ht, unsigned long size, static long nr_cpus_mask = -1; static long split_count_mask = -1; +static int split_count_order = -1; #if defined(HAVE_SYSCONF) static void ht_init_nr_cpus_mask(void) @@ -606,6 +601,8 @@ void alloc_split_items_count(struct cds_lfht *ht) split_count_mask = DEFAULT_SPLIT_COUNT_MASK; else split_count_mask = nr_cpus_mask; + split_count_order = + cds_lfht_get_count_order_ulong(split_count_mask + 1); } assert(split_count_mask >= 0); @@ -624,7 +621,7 @@ void free_split_items_count(struct cds_lfht *ht) poison_free(ht->split_count); } -#if defined(HAVE_SCHED_GETCPU) +#if defined(HAVE_SCHED_GETCPU) && !defined(VALGRIND) static int ht_get_split_count_index(unsigned long hash) { @@ -721,14 +718,39 @@ void check_resize(struct cds_lfht *ht, unsigned long size, uint32_t chain_len) * Use bucket-local length for small table expand and for * environments lacking per-cpu data support. */ - if (count >= (1UL << COUNT_COMMIT_ORDER)) + if (count >= (1UL << (COUNT_COMMIT_ORDER + split_count_order))) return; if (chain_len > 100) dbg_printf("WARNING: large chain length: %u.\n", chain_len); - if (chain_len >= CHAIN_LEN_RESIZE_THRESHOLD) - cds_lfht_resize_lazy_grow(ht, size, - cds_lfht_get_count_order_u32(chain_len - (CHAIN_LEN_TARGET - 1))); + if (chain_len >= CHAIN_LEN_RESIZE_THRESHOLD) { + int growth; + + /* + * Ideal growth calculated based on chain length. + */ + growth = cds_lfht_get_count_order_u32(chain_len + - (CHAIN_LEN_TARGET - 1)); + if ((ht->flags & CDS_LFHT_ACCOUNTING) + && (size << growth) + >= (1UL << (COUNT_COMMIT_ORDER + + split_count_order))) { + /* + * If ideal growth expands the hash table size + * beyond the "small hash table" sizes, use the + * maximum small hash table size to attempt + * expanding the hash table. This only applies + * when node accounting is available, otherwise + * the chain length is used to expand the hash + * table in every case. + */ + growth = COUNT_COMMIT_ORDER + split_count_order + - cds_lfht_get_count_order_ulong(size); + if (growth <= 0) + return; + } + cds_lfht_resize_lazy_grow(ht, size, growth); + } } static @@ -1733,12 +1755,27 @@ int cds_lfht_delete_bucket(struct cds_lfht *ht) int cds_lfht_destroy(struct cds_lfht *ht, pthread_attr_t **attr) { int ret; +#ifdef rcu_read_ongoing_mb + int was_online; +#endif /* Wait for in-flight resize operations to complete */ _CMM_STORE_SHARED(ht->in_progress_destroy, 1); cmm_smp_mb(); /* Store destroy before load resize */ +#ifdef rcu_read_ongoing_mb + was_online = ht->flavor->read_ongoing(); + if (was_online) + ht->flavor->thread_offline(); + /* Calling with RCU read-side held is an error. */ + if (ht->flavor->read_ongoing()) { + ret = -EINVAL; + if (was_online) + ht->flavor->thread_online(); + goto end; + } +#endif while (uatomic_read(&ht->in_progress_resize)) - poll(NULL, 0, 100); /* wait for 100ms */ + (void) poll(NULL, 0, 100); /* wait for 100ms */ ret = cds_lfht_delete_bucket(ht); if (ret) return ret; @@ -1746,6 +1783,9 @@ int cds_lfht_destroy(struct cds_lfht *ht, pthread_attr_t **attr) if (attr) *attr = ht->resize_attr; poison_free(ht); +#ifdef rcu_read_ongoing_mb +end: +#endif return ret; } @@ -1874,13 +1914,34 @@ void resize_target_update_count(struct cds_lfht *ht, void cds_lfht_resize(struct cds_lfht *ht, unsigned long new_size) { +#ifdef rcu_read_ongoing_mb + int was_online; + + was_online = ht->flavor->read_ongoing(); + if (was_online) + ht->flavor->thread_offline(); + /* Calling with RCU read-side held is an error. */ + if (ht->flavor->read_ongoing()) { + static int print_once; + + if (!CMM_LOAD_SHARED(print_once)) + fprintf(stderr, "[error] rculfhash: cds_lfht_resize " + "called with RCU read-side lock held.\n"); + CMM_STORE_SHARED(print_once, 1); + assert(0); + goto end; + } +#endif resize_target_update_count(ht, new_size); CMM_STORE_SHARED(ht->resize_initiated, 1); - ht->flavor->thread_offline(); pthread_mutex_lock(&ht->resize_mutex); _do_cds_lfht_resize(ht); pthread_mutex_unlock(&ht->resize_mutex); - ht->flavor->thread_online(); +#ifdef rcu_read_ongoing_mb +end: + if (was_online) + ht->flavor->thread_online(); +#endif } static @@ -1914,7 +1975,7 @@ void __cds_lfht_resize_lazy_launch(struct cds_lfht *ht) uatomic_dec(&ht->in_progress_resize); return; } - work = malloc(sizeof(*work)); + work = zmalloc(sizeof(*work)); if (work == NULL) { dbg_printf("error allocating resize work, bailing out\n"); uatomic_dec(&ht->in_progress_resize);