X-Git-Url: http://git.efficios.com/?p=lttng-tools.git;a=blobdiff_plain;f=src%2Fcommon%2Fconsumer-timer.c;h=5764b13339cafd0366d68e7120519c4333ef3057;hp=f2dff4cd980aae5258cfc19a3c268bd800c955c8;hb=ea26306076d26c40ed31e4f9170dc2852bda502f;hpb=d98a47c754f867bee113325ff69419f0829de0bb diff --git a/src/common/consumer-timer.c b/src/common/consumer-timer.c index f2dff4cd9..5764b1333 100644 --- a/src/common/consumer-timer.c +++ b/src/common/consumer-timer.c @@ -16,14 +16,20 @@ * Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. */ -#define _GNU_SOURCE +#define _LGPL_SOURCE #include #include #include +#include #include +#include +#include +#include +#include #include "consumer-timer.h" +#include "consumer-testpoint.h" #include "ust-consumer/ust-consumer.h" static struct timer_signal_data timer_signal = { @@ -46,11 +52,15 @@ static void setmask(sigset_t *mask) } ret = sigaddset(mask, LTTNG_CONSUMER_SIG_SWITCH); if (ret) { - PERROR("sigaddset"); + PERROR("sigaddset switch"); } ret = sigaddset(mask, LTTNG_CONSUMER_SIG_TEARDOWN); if (ret) { - PERROR("sigaddset"); + PERROR("sigaddset teardown"); + } + ret = sigaddset(mask, LTTNG_CONSUMER_SIG_LIVE); + if (ret) { + PERROR("sigaddset live"); } } @@ -78,7 +88,21 @@ static void metadata_switch_timer(struct lttng_consumer_local_data *ctx, switch (ctx->type) { case LTTNG_CONSUMER32_UST: case LTTNG_CONSUMER64_UST: - ret = lttng_ustconsumer_request_metadata(ctx, channel); + /* + * Locks taken by lttng_ustconsumer_request_metadata(): + * - metadata_socket_lock + * - Calling lttng_ustconsumer_recv_metadata(): + * - channel->metadata_cache->lock + * - Calling consumer_metadata_cache_flushed(): + * - channel->timer_lock + * - channel->metadata_cache->lock + * + * Ensure that neither consumer_data.lock nor + * channel->lock are taken within this function, since + * they are held while consumer_timer_switch_stop() is + * called. + */ + ret = lttng_ustconsumer_request_metadata(ctx, channel, 1, 1); if (ret < 0) { channel->switch_timer_error = 1; } @@ -90,6 +114,252 @@ static void metadata_switch_timer(struct lttng_consumer_local_data *ctx, } } +static int send_empty_index(struct lttng_consumer_stream *stream, uint64_t ts, + uint64_t stream_id) +{ + int ret; + struct ctf_packet_index index; + + memset(&index, 0, sizeof(index)); + index.stream_id = htobe64(stream_id); + index.timestamp_end = htobe64(ts); + ret = consumer_stream_write_index(stream, &index); + if (ret < 0) { + goto error; + } + +error: + return ret; +} + +int consumer_flush_kernel_index(struct lttng_consumer_stream *stream) +{ + uint64_t ts, stream_id; + int ret; + + ret = kernctl_get_current_timestamp(stream->wait_fd, &ts); + if (ret < 0) { + ERR("Failed to get the current timestamp"); + goto end; + } + ret = kernctl_buffer_flush(stream->wait_fd); + if (ret < 0) { + ERR("Failed to flush kernel stream"); + goto end; + } + ret = kernctl_snapshot(stream->wait_fd); + if (ret < 0) { + if (errno != EAGAIN && errno != ENODATA) { + PERROR("live timer kernel snapshot"); + ret = -1; + goto end; + } + ret = kernctl_get_stream_id(stream->wait_fd, &stream_id); + if (ret < 0) { + PERROR("kernctl_get_stream_id"); + goto end; + } + DBG("Stream %" PRIu64 " empty, sending beacon", stream->key); + ret = send_empty_index(stream, ts, stream_id); + if (ret < 0) { + goto end; + } + } + ret = 0; +end: + return ret; +} + +static int check_kernel_stream(struct lttng_consumer_stream *stream) +{ + int ret; + + /* + * While holding the stream mutex, try to take a snapshot, if it + * succeeds, it means that data is ready to be sent, just let the data + * thread handle that. Otherwise, if the snapshot returns EAGAIN, it + * means that there is no data to read after the flush, so we can + * safely send the empty index. + * + * Doing a trylock and checking if waiting on metadata if + * trylock fails. Bail out of the stream is indeed waiting for + * metadata to be pushed. Busy wait on trylock otherwise. + */ + for (;;) { + ret = pthread_mutex_trylock(&stream->lock); + switch (ret) { + case 0: + break; /* We have the lock. */ + case EBUSY: + pthread_mutex_lock(&stream->metadata_timer_lock); + if (stream->waiting_on_metadata) { + ret = 0; + stream->missed_metadata_flush = true; + pthread_mutex_unlock(&stream->metadata_timer_lock); + goto end; /* Bail out. */ + } + pthread_mutex_unlock(&stream->metadata_timer_lock); + /* Try again. */ + caa_cpu_relax(); + continue; + default: + ERR("Unexpected pthread_mutex_trylock error %d", ret); + ret = -1; + goto end; + } + break; + } + ret = consumer_flush_kernel_index(stream); + pthread_mutex_unlock(&stream->lock); +end: + return ret; +} + +int consumer_flush_ust_index(struct lttng_consumer_stream *stream) +{ + uint64_t ts, stream_id; + int ret; + + ret = cds_lfht_is_node_deleted(&stream->node.node); + if (ret) { + goto end; + } + + ret = lttng_ustconsumer_get_current_timestamp(stream, &ts); + if (ret < 0) { + ERR("Failed to get the current timestamp"); + goto end; + } + lttng_ustconsumer_flush_buffer(stream, 1); + ret = lttng_ustconsumer_take_snapshot(stream); + if (ret < 0) { + if (ret != -EAGAIN) { + ERR("Taking UST snapshot"); + ret = -1; + goto end; + } + ret = lttng_ustconsumer_get_stream_id(stream, &stream_id); + if (ret < 0) { + PERROR("ustctl_get_stream_id"); + goto end; + } + DBG("Stream %" PRIu64 " empty, sending beacon", stream->key); + ret = send_empty_index(stream, ts, stream_id); + if (ret < 0) { + goto end; + } + } + ret = 0; +end: + return ret; +} + +static int check_ust_stream(struct lttng_consumer_stream *stream) +{ + int ret; + + assert(stream); + assert(stream->ustream); + /* + * While holding the stream mutex, try to take a snapshot, if it + * succeeds, it means that data is ready to be sent, just let the data + * thread handle that. Otherwise, if the snapshot returns EAGAIN, it + * means that there is no data to read after the flush, so we can + * safely send the empty index. + * + * Doing a trylock and checking if waiting on metadata if + * trylock fails. Bail out of the stream is indeed waiting for + * metadata to be pushed. Busy wait on trylock otherwise. + */ + for (;;) { + ret = pthread_mutex_trylock(&stream->lock); + switch (ret) { + case 0: + break; /* We have the lock. */ + case EBUSY: + pthread_mutex_lock(&stream->metadata_timer_lock); + if (stream->waiting_on_metadata) { + ret = 0; + stream->missed_metadata_flush = true; + pthread_mutex_unlock(&stream->metadata_timer_lock); + goto end; /* Bail out. */ + } + pthread_mutex_unlock(&stream->metadata_timer_lock); + /* Try again. */ + caa_cpu_relax(); + continue; + default: + ERR("Unexpected pthread_mutex_trylock error %d", ret); + ret = -1; + goto end; + } + break; + } + ret = consumer_flush_ust_index(stream); + pthread_mutex_unlock(&stream->lock); +end: + return ret; +} + +/* + * Execute action on a live timer + */ +static void live_timer(struct lttng_consumer_local_data *ctx, + int sig, siginfo_t *si, void *uc) +{ + int ret; + struct lttng_consumer_channel *channel; + struct lttng_consumer_stream *stream; + struct lttng_ht *ht; + struct lttng_ht_iter iter; + + channel = si->si_value.sival_ptr; + assert(channel); + + if (channel->switch_timer_error) { + goto error; + } + ht = consumer_data.stream_per_chan_id_ht; + + DBG("Live timer for channel %" PRIu64, channel->key); + + rcu_read_lock(); + switch (ctx->type) { + case LTTNG_CONSUMER32_UST: + case LTTNG_CONSUMER64_UST: + cds_lfht_for_each_entry_duplicate(ht->ht, + ht->hash_fct(&channel->key, lttng_ht_seed), + ht->match_fct, &channel->key, &iter.iter, + stream, node_channel_id.node) { + ret = check_ust_stream(stream); + if (ret < 0) { + goto error_unlock; + } + } + break; + case LTTNG_CONSUMER_KERNEL: + cds_lfht_for_each_entry_duplicate(ht->ht, + ht->hash_fct(&channel->key, lttng_ht_seed), + ht->match_fct, &channel->key, &iter.iter, + stream, node_channel_id.node) { + ret = check_kernel_stream(stream); + if (ret < 0) { + goto error_unlock; + } + } + break; + case LTTNG_CONSUMER_UNKNOWN: + assert(0); + break; + } + +error_unlock: + rcu_read_unlock(); + +error: + return; +} + static void consumer_timer_signal_thread_qs(unsigned int signr) { @@ -198,11 +468,68 @@ void consumer_timer_switch_stop(struct lttng_consumer_channel *channel) channel->switch_timer_enabled = 0; } +/* + * Set the timer for the live mode. + */ +void consumer_timer_live_start(struct lttng_consumer_channel *channel, + int live_timer_interval) +{ + int ret; + struct sigevent sev; + struct itimerspec its; + + assert(channel); + assert(channel->key); + + if (live_timer_interval <= 0) { + return; + } + + sev.sigev_notify = SIGEV_SIGNAL; + sev.sigev_signo = LTTNG_CONSUMER_SIG_LIVE; + sev.sigev_value.sival_ptr = channel; + ret = timer_create(CLOCKID, &sev, &channel->live_timer); + if (ret == -1) { + PERROR("timer_create"); + } + channel->live_timer_enabled = 1; + + its.it_value.tv_sec = live_timer_interval / 1000000; + its.it_value.tv_nsec = live_timer_interval % 1000000; + its.it_interval.tv_sec = its.it_value.tv_sec; + its.it_interval.tv_nsec = its.it_value.tv_nsec; + + ret = timer_settime(channel->live_timer, 0, &its, NULL); + if (ret == -1) { + PERROR("timer_settime"); + } +} + +/* + * Stop and delete timer. + */ +void consumer_timer_live_stop(struct lttng_consumer_channel *channel) +{ + int ret; + + assert(channel); + + ret = timer_delete(channel->live_timer); + if (ret == -1) { + PERROR("timer_delete"); + } + + consumer_timer_signal_thread_qs(LTTNG_CONSUMER_SIG_LIVE); + + channel->live_timer = 0; + channel->live_timer_enabled = 0; +} + /* * Block the RT signals for the entire process. It must be called from the * consumer main before creating the threads */ -void consumer_signal_init(void) +int consumer_signal_init(void) { int ret; sigset_t mask; @@ -213,27 +540,42 @@ void consumer_signal_init(void) if (ret) { errno = ret; PERROR("pthread_sigmask"); + return -1; } + return 0; } /* - * This thread is the sighandler for signals LTTNG_CONSUMER_SIG_SWITCH and - * LTTNG_CONSUMER_SIG_TEARDOWN that are emitted by the periodic timer to check - * if new metadata is available. + * This thread is the sighandler for signals LTTNG_CONSUMER_SIG_SWITCH, + * LTTNG_CONSUMER_SIG_TEARDOWN and LTTNG_CONSUMER_SIG_LIVE. */ -void *consumer_timer_metadata_thread(void *data) +void *consumer_timer_thread(void *data) { int signr; sigset_t mask; siginfo_t info; struct lttng_consumer_local_data *ctx = data; + rcu_register_thread(); + + health_register(health_consumerd, HEALTH_CONSUMERD_TYPE_METADATA_TIMER); + + if (testpoint(consumerd_thread_metadata_timer)) { + goto error_testpoint; + } + + health_code_update(); + /* Only self thread will receive signal mask. */ setmask(&mask); CMM_STORE_SHARED(timer_signal.tid, pthread_self()); while (1) { + health_code_update(); + + health_poll_entry(); signr = sigwaitinfo(&mask, &info); + health_poll_exit(); if (signr == -1) { if (errno != EINTR) { PERROR("sigwaitinfo"); @@ -246,10 +588,20 @@ void *consumer_timer_metadata_thread(void *data) CMM_STORE_SHARED(timer_signal.qs_done, 1); cmm_smp_mb(); DBG("Signal timer metadata thread teardown"); + } else if (signr == LTTNG_CONSUMER_SIG_LIVE) { + live_timer(ctx, info.si_signo, &info, NULL); } else { ERR("Unexpected signal %d\n", info.si_signo); } } +error_testpoint: + /* Only reached in testpoint error */ + health_error(); + health_unregister(health_consumerd); + + rcu_unregister_thread(); + + /* Never return */ return NULL; }