X-Git-Url: http://git.efficios.com/?p=lttng-tools.git;a=blobdiff_plain;f=ltt-sessiond%2Fmain.c;h=d2240d9dd89a1f9cec011e233955b9c33a441b7e;hp=edbdd643a84f5e1c69b3118663af7fafb2b274fa;hb=e848fc768a950acc1b823fcb139057a4fac1442a;hpb=f9815039b2c793170f6c411b3949aaa76d5b9141 diff --git a/ltt-sessiond/main.c b/ltt-sessiond/main.c index edbdd643a..d2240d9dd 100644 --- a/ltt-sessiond/main.c +++ b/ltt-sessiond/main.c @@ -1,6 +1,6 @@ /* * Copyright (C) 2011 - David Goulet - * Copyright (C) 2011 - Mathieu Desnoyers + * Mathieu Desnoyers * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License @@ -22,18 +22,20 @@ #include #include #include -#include #include #include #include #include #include #include +#include #include #include #include #include #include +#include +#include #include #include @@ -41,14 +43,16 @@ #include #include +#include "compat/poll.h" #include "context.h" #include "futex.h" #include "kernel-ctl.h" #include "ltt-sessiond.h" +#include "shm.h" #include "traceable-app.h" #include "ust-ctl.h" #include "utils.h" -#include "ust-comm.h" +#include "ust-ctl.h" /* Const values */ const char default_home_dir[] = DEFAULT_HOME_DIR; @@ -68,13 +72,13 @@ static int opt_daemon; static int is_root; /* Set to 1 if the daemon is running as root */ static pid_t ppid; /* Parent PID for --sig-parent option */ static pid_t kconsumerd_pid; -static struct pollfd *kernel_pollfd; static int dispatch_thread_exit; static char apps_unix_sock_path[PATH_MAX]; /* Global application Unix socket path */ static char client_unix_sock_path[PATH_MAX]; /* Global client Unix socket path */ static char kconsumerd_err_unix_sock_path[PATH_MAX]; /* kconsumerd error Unix socket path */ static char kconsumerd_cmd_unix_sock_path[PATH_MAX]; /* kconsumerd command Unix socket path */ +static char wait_shm_path[PATH_MAX]; /* global wait shm path for UST */ /* Sockets and FDs */ static int client_sock; @@ -105,9 +109,9 @@ static pthread_t kernel_thread; static pthread_t dispatch_thread; static sem_t kconsumerd_sem; -static pthread_mutex_t kconsumerd_pid_mutex; /* Mutex to control kconsumerd pid assignation */ -static int modprobe_remove_kernel_modules(void); +/* Mutex to control kconsumerd pid assignation */ +static pthread_mutex_t kconsumerd_pid_mutex; /* * UST registration command queue. This queue is tied with a futex and uses a N @@ -130,6 +134,88 @@ static struct ust_cmd_queue ust_cmd_queue; */ static struct ltt_session_list *session_list_ptr; +/* + * Create a poll set with O_CLOEXEC and add the thread quit pipe to the set. + */ +static int create_thread_poll_set(struct lttng_poll_event *events, + unsigned int size) +{ + int ret; + + if (events == NULL || size == 0) { + ret = -1; + goto error; + } + + ret = lttng_poll_create(events, size, LTTNG_CLOEXEC); + if (ret < 0) { + goto error; + } + + /* Add quit pipe */ + ret = lttng_poll_add(events, thread_quit_pipe[0], LPOLLIN); + if (ret < 0) { + goto error; + } + + return 0; + +error: + return ret; +} + +/* + * Check if the thread quit pipe was triggered. + * + * Return 1 if it was triggered else 0; + */ +static int check_thread_quit_pipe(int fd, uint32_t events) +{ + if (fd == thread_quit_pipe[0] && (events & LPOLLIN)) { + return 1; + } + + return 0; +} + +/* + * Remove modules in reverse load order. + */ +static int modprobe_remove_kernel_modules(void) +{ + int ret = 0, i; + char modprobe[256]; + + for (i = ARRAY_SIZE(kernel_modules_list) - 1; i >= 0; i--) { + ret = snprintf(modprobe, sizeof(modprobe), + "/sbin/modprobe --remove --quiet %s", + kernel_modules_list[i].name); + if (ret < 0) { + perror("snprintf modprobe --remove"); + goto error; + } + modprobe[sizeof(modprobe) - 1] = '\0'; + ret = system(modprobe); + if (ret == -1) { + ERR("Unable to launch modprobe --remove for module %s", + kernel_modules_list[i].name); + } else if (kernel_modules_list[i].required + && WEXITSTATUS(ret) != 0) { + ERR("Unable to remove module %s", + kernel_modules_list[i].name); + } else { + DBG("Modprobe removal successful %s", + kernel_modules_list[i].name); + } + } + +error: + return ret; +} + +/* + * Return group ID of the tracing group or -1 if not found. + */ static gid_t allowed_group(void) { struct group *grp; @@ -147,7 +233,7 @@ static gid_t allowed_group(void) } /* - * Init quit pipe. + * Init thread quit pipe. * * Return -1 on error or 0 if all pipes are created. */ @@ -193,10 +279,15 @@ static void teardown_kernel_session(struct ltt_session *session) */ static void stop_threads(void) { + int ret; + /* Stopping all threads */ DBG("Terminating all threads"); - close(thread_quit_pipe[0]); - close(thread_quit_pipe[1]); + ret = write(thread_quit_pipe[1], "!", 1); + if (ret < 0) { + ERR("write error on thread quit pipe"); + } + /* Dispatch thread */ dispatch_thread_exit = 1; futex_nto1_wake(&ust_cmd_queue.futex); @@ -214,21 +305,23 @@ static void cleanup(void) DBG("Cleaning up"); /* */ - MSG("\n%c[%d;%dm*** assert failed *** ==> %c[%dm%c[%d;%dm" + MSG("%c[%d;%dm*** assert failed *** ==> %c[%dm%c[%d;%dm" "Matthew, BEET driven development works!%c[%dm", 27, 1, 31, 27, 0, 27, 1, 33, 27, 0); /* */ - DBG("Removing %s directory", LTTNG_RUNDIR); - ret = asprintf(&cmd, "rm -rf " LTTNG_RUNDIR); - if (ret < 0) { - ERR("asprintf failed. Something is really wrong!"); - } + if (is_root) { + DBG("Removing %s directory", LTTNG_RUNDIR); + ret = asprintf(&cmd, "rm -rf " LTTNG_RUNDIR); + if (ret < 0) { + ERR("asprintf failed. Something is really wrong!"); + } - /* Remove lttng run directory */ - ret = system(cmd); - if (ret < 0) { - ERR("Unable to clean " LTTNG_RUNDIR); + /* Remove lttng run directory */ + ret = system(cmd); + if (ret < 0) { + ERR("Unable to clean " LTTNG_RUNDIR); + } } DBG("Cleaning up all session"); @@ -252,8 +345,13 @@ static void cleanup(void) DBG("Closing kernel fd"); close(kernel_tracer_fd); - DBG("Unloading kernel modules"); - modprobe_remove_kernel_modules(); + if (is_root) { + DBG("Unloading kernel modules"); + modprobe_remove_kernel_modules(); + } + + close(thread_quit_pipe[0]); + close(thread_quit_pipe[1]); } /* @@ -403,65 +501,30 @@ error: return ret; } -#ifdef DISABLED /* - * Return a socket connected to the libust communication socket of the - * application identified by the pid. - * - * If the pid is not found in the traceable list, return -1 to indicate error. + * Notify UST applications using the shm mmap futex. */ -static int ust_connect_app(pid_t pid) +static int notify_ust_apps(int active) { - int sock; - struct ltt_traceable_app *lta; + char *wait_shm_mmap; - DBG("Connect to application pid %d", pid); + DBG("Notifying applications of session daemon state: %d", active); - lta = find_app_by_pid(pid); - if (lta == NULL) { - /* App not found */ - DBG("Application pid %d not found", pid); - return -1; - } - - sock = ustctl_connect_pid(lta->pid); - if (sock < 0) { - ERR("Fail connecting to the PID %d", pid); - } - - return sock; -} - -/* - * Notify apps by writing 42 to a named pipe using name. Every applications - * waiting for a ltt-sessiond will be notified and re-register automatically to - * the session daemon. - * - * Return open or write error value. - */ -static int notify_apps(const char *name) -{ - int fd; - int ret = -1; - - DBG("Notify the global application pipe"); - - /* Try opening the global pipe */ - fd = open(name, O_WRONLY); - if (fd < 0) { + /* See shm.c for this call implying mmap, shm and futex calls */ + wait_shm_mmap = shm_ust_get_mmap(wait_shm_path, is_root); + if (wait_shm_mmap == NULL) { goto error; } - /* Notify by writing on the pipe */ - ret = write(fd, "42", 2); - if (ret < 0) { - perror("write"); - } + /* Wake waiting process */ + futex_wait_update((int32_t *) wait_shm_mmap, active); + + /* Apps notified successfully */ + return 0; error: - return ret; + return -1; } -#endif /* DISABLED */ /* * Setup the outgoing data buffer for the response (llm) by allocating the @@ -497,23 +560,17 @@ error: } /* - * Update the kernel pollfd set of all channel fd available over all tracing + * Update the kernel poll set of all channel fd available over all tracing * session. Add the wakeup pipe at the end of the set. */ -static int update_kernel_pollfd(void) +static int update_kernel_poll(struct lttng_poll_event *events) { - int i = 0; - /* - * The wakup pipe and the quit pipe are needed so the number of fds starts - * at 2 for those pipes. - */ - unsigned int nb_fd = 2; + int ret; struct ltt_session *session; struct ltt_kernel_channel *channel; - DBG("Updating kernel_pollfd"); + DBG("Updating kernel poll set"); - /* Get the number of channel of all kernel session */ lock_session_list(); cds_list_for_each_entry(session, &session_list_ptr->head, list) { lock_session(session); @@ -521,46 +578,21 @@ static int update_kernel_pollfd(void) unlock_session(session); continue; } - nb_fd += session->kernel_session->channel_count; - unlock_session(session); - } - - DBG("Resizing kernel_pollfd to size %d", nb_fd); - - kernel_pollfd = realloc(kernel_pollfd, nb_fd * sizeof(struct pollfd)); - if (kernel_pollfd == NULL) { - perror("malloc kernel_pollfd"); - goto error; - } - cds_list_for_each_entry(session, &session_list_ptr->head, list) { - lock_session(session); - if (session->kernel_session == NULL) { - unlock_session(session); - continue; - } - if (i >= nb_fd) { - ERR("To much channel for kernel_pollfd size"); - unlock_session(session); - break; - } cds_list_for_each_entry(channel, &session->kernel_session->channel_list.head, list) { - kernel_pollfd[i].fd = channel->fd; - kernel_pollfd[i].events = POLLIN | POLLRDNORM; - i++; + /* Add channel fd to the kernel poll set */ + ret = lttng_poll_add(events, channel->fd, LPOLLIN | LPOLLRDNORM); + if (ret < 0) { + unlock_session(session); + goto error; + } + DBG("Channel fd %d added to kernel set", channel->fd); } unlock_session(session); } unlock_session_list(); - /* Adding wake up pipe */ - kernel_pollfd[nb_fd - 2].fd = kernel_poll_pipe[0]; - kernel_pollfd[nb_fd - 2].events = POLLIN; - - /* Adding the quit pipe */ - kernel_pollfd[nb_fd - 1].fd = thread_quit_pipe[0]; - - return nb_fd; + return 0; error: unlock_session_list(); @@ -594,36 +626,38 @@ static int update_kernel_stream(int fd) session->kernel_session->consumer_fd = kconsumerd_cmd_sock; } - cds_list_for_each_entry(channel, &session->kernel_session->channel_list.head, list) { + cds_list_for_each_entry(channel, + &session->kernel_session->channel_list.head, list) { if (channel->fd == fd) { DBG("Channel found, updating kernel streams"); ret = kernel_open_channel_stream(channel); if (ret < 0) { - goto end; + goto error; } /* - * Have we already sent fds to the consumer? If yes, it means that - * tracing is started so it is safe to send our updated stream fds. + * Have we already sent fds to the consumer? If yes, it means + * that tracing is started so it is safe to send our updated + * stream fds. */ if (session->kernel_session->kconsumer_fds_sent == 1) { - ret = send_kconsumerd_channel_fds(session->kernel_session->consumer_fd, - channel); + ret = send_kconsumerd_channel_fds( + session->kernel_session->consumer_fd, channel); if (ret < 0) { - goto end; + goto error; } } - goto end; + goto error; } } unlock_session(session); } + unlock_session_list(); + return ret; -end: +error: + unlock_session(session); unlock_session_list(); - if (session) { - unlock_session(session); - } return ret; } @@ -635,79 +669,93 @@ end: */ static void *thread_manage_kernel(void *data) { - int ret, i, nb_fd = 0; + int ret, i, pollfd, update_poll_flag = 1; + uint32_t revents, nb_fd; char tmp; - int update_poll_flag = 1; + struct lttng_poll_event events; DBG("Thread manage kernel started"); + ret = create_thread_poll_set(&events, 2); + if (ret < 0) { + goto error; + } + + ret = lttng_poll_add(&events, kernel_poll_pipe[0], LPOLLIN); + if (ret < 0) { + goto error; + } + while (1) { if (update_poll_flag == 1) { - nb_fd = update_kernel_pollfd(); - if (nb_fd < 0) { + ret = update_kernel_poll(&events); + if (ret < 0) { goto error; } update_poll_flag = 0; } - DBG("Polling on %d fds", nb_fd); + nb_fd = LTTNG_POLL_GETNB(&events); + + DBG("Thread kernel polling on %d fds", nb_fd); + + /* Zeroed the poll events */ + lttng_poll_reset(&events); /* Poll infinite value of time */ - ret = poll(kernel_pollfd, nb_fd, -1); + ret = lttng_poll_wait(&events, -1); if (ret < 0) { - perror("poll kernel thread"); goto error; } else if (ret == 0) { /* Should not happen since timeout is infinite */ + ERR("Return value of poll is 0 with an infinite timeout.\n" + "This should not have happened! Continuing..."); continue; } - /* Thread quit pipe has been closed. Killing thread. */ - if (kernel_pollfd[nb_fd - 1].revents == POLLNVAL) { - goto error; - } - - DBG("Kernel poll event triggered"); + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); - /* - * Check if the wake up pipe was triggered. If so, the kernel_pollfd - * must be updated. - */ - switch (kernel_pollfd[nb_fd - 2].revents) { - case POLLIN: - ret = read(kernel_poll_pipe[0], &tmp, 1); - update_poll_flag = 1; - continue; - case POLLERR: - goto error; - default: - break; - } + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { + goto error; + } - for (i = 0; i < nb_fd; i++) { - switch (kernel_pollfd[i].revents) { - /* - * New CPU detected by the kernel. Adding kernel stream to kernel - * session and updating the kernel consumer - */ - case POLLIN | POLLRDNORM: - ret = update_kernel_stream(kernel_pollfd[i].fd); - if (ret < 0) { - continue; + /* Check for data on kernel pipe */ + if (pollfd == kernel_poll_pipe[0] && (revents & LPOLLIN)) { + ret = read(kernel_poll_pipe[0], &tmp, 1); + update_poll_flag = 1; + continue; + } else { + /* + * New CPU detected by the kernel. Adding kernel stream to + * kernel session and updating the kernel consumer + */ + if (revents & LPOLLIN) { + ret = update_kernel_stream(pollfd); + if (ret < 0) { + continue; + } + break; + /* + * TODO: We might want to handle the LPOLLERR | LPOLLHUP + * and unregister kernel stream at this point. + */ } - break; } } } error: DBG("Kernel thread dying"); - if (kernel_pollfd) { - free(kernel_pollfd); - } - close(kernel_poll_pipe[0]); close(kernel_poll_pipe[1]); + + lttng_poll_clean(&events); + return NULL; } @@ -716,9 +764,10 @@ error: */ static void *thread_manage_kconsumerd(void *data) { - int sock = 0, ret; + int sock = 0, i, ret, pollfd; + uint32_t revents, nb_fd; enum lttcomm_return_code code; - struct pollfd pollfd[2]; + struct lttng_poll_event events; DBG("[thread] Manage kconsumerd started"); @@ -727,26 +776,46 @@ static void *thread_manage_kconsumerd(void *data) goto error; } - /* First fd is always the quit pipe */ - pollfd[0].fd = thread_quit_pipe[0]; + /* + * Pass 2 as size here for the thread quit pipe and kconsumerd_err_sock. + * Nothing more will be added to this poll set. + */ + ret = create_thread_poll_set(&events, 2); + if (ret < 0) { + goto error; + } - /* Apps socket */ - pollfd[1].fd = kconsumerd_err_sock; - pollfd[1].events = POLLIN; + ret = lttng_poll_add(&events, kconsumerd_err_sock, LPOLLIN | LPOLLRDHUP); + if (ret < 0) { + goto error; + } + + nb_fd = LTTNG_POLL_GETNB(&events); /* Inifinite blocking call, waiting for transmission */ - ret = poll(pollfd, 2, -1); + ret = lttng_poll_wait(&events, -1); if (ret < 0) { - perror("poll kconsumerd thread"); goto error; } - /* Thread quit pipe has been closed. Killing thread. */ - if (pollfd[0].revents == POLLNVAL) { - goto error; - } else if (pollfd[1].revents == POLLERR) { - ERR("Kconsumerd err socket poll error"); - goto error; + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); + + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { + goto error; + } + + /* Event on the registration socket */ + if (pollfd == kconsumerd_err_sock) { + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + ERR("Kconsumerd err socket poll error"); + goto error; + } + } } sock = lttcomm_accept_unix_sock(kconsumerd_err_sock); @@ -776,25 +845,46 @@ static void *thread_manage_kconsumerd(void *data) goto error; } - /* Kconsumerd err socket */ - pollfd[1].fd = sock; - pollfd[1].events = POLLIN; - - /* Inifinite blocking call, waiting for transmission */ - ret = poll(pollfd, 2, -1); + /* Remove the kconsumerd error socket since we have established a connexion */ + ret = lttng_poll_del(&events, kconsumerd_err_sock); if (ret < 0) { - perror("poll kconsumerd thread"); goto error; } - /* Thread quit pipe has been closed. Killing thread. */ - if (pollfd[0].revents == POLLNVAL) { + ret = lttng_poll_add(&events, sock, LPOLLIN | LPOLLRDHUP); + if (ret < 0) { goto error; - } else if (pollfd[1].revents == POLLERR) { - ERR("Kconsumerd err socket second poll error"); + } + + /* Update number of fd */ + nb_fd = LTTNG_POLL_GETNB(&events); + + /* Inifinite blocking call, waiting for transmission */ + ret = lttng_poll_wait(&events, -1); + if (ret < 0) { goto error; } + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); + + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { + goto error; + } + + /* Event on the kconsumerd socket */ + if (pollfd == sock) { + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + ERR("Kconsumerd err socket second poll error"); + goto error; + } + } + } + /* Wait for any kconsumerd error */ ret = lttcomm_recv_unix_sock(sock, &code, sizeof(enum lttcomm_return_code)); if (ret <= 0) { @@ -806,69 +896,17 @@ static void *thread_manage_kconsumerd(void *data) error: DBG("Kconsumerd thread dying"); - if (kconsumerd_err_sock) { - close(kconsumerd_err_sock); - } - if (kconsumerd_cmd_sock) { - close(kconsumerd_cmd_sock); - } - if (sock) { - close(sock); - } + close(kconsumerd_err_sock); + close(kconsumerd_cmd_sock); + close(sock); unlink(kconsumerd_err_unix_sock_path); unlink(kconsumerd_cmd_unix_sock_path); - kconsumerd_pid = 0; - return NULL; -} -/* - * Reallocate the apps command pollfd structure of nb_fd size. - * - * The first two fds must be there at all time. - */ -static int update_apps_cmd_pollfd(unsigned int nb_fd, struct pollfd **pollfd) -{ - /* Can't accept pollfd less than 2 */ - if (nb_fd < 2) { - goto end; - } - - *pollfd = realloc(*pollfd, nb_fd * sizeof(struct pollfd)); - if (*pollfd == NULL) { - perror("realloc manage apps pollfd"); - goto error; - } - - /* First fd is always the quit pipe */ - (*pollfd)[0].fd = thread_quit_pipe[0]; - /* Apps command pipe */ - (*pollfd)[1].fd = apps_cmd_pipe[0]; - (*pollfd)[1].events = POLLIN; - - DBG("Apps cmd pollfd realloc of size %d", nb_fd); - -end: - return 0; - -error: - return -1; -} - -/* - * Send registration done packet to the application. - */ -static int send_ust_register_done(int sock) -{ - struct lttcomm_ust_msg lum; - - DBG("Sending register done command to %d", sock); + lttng_poll_clean(&events); - lum.cmd = LTTNG_UST_REGISTER_DONE; - lum.handle = LTTNG_UST_ROOT_HANDLE; - - return ustcomm_send_command(sock, &lum); + return NULL; } /* @@ -876,94 +914,107 @@ static int send_ust_register_done(int sock) */ static void *thread_manage_apps(void *data) { - int i, ret, count; - unsigned int nb_fd = 2; - int update_poll_flag = 1; - struct pollfd *pollfd = NULL; + int i, ret, pollfd; + uint32_t revents, nb_fd; struct ust_command ust_cmd; + struct lttng_poll_event events; DBG("[thread] Manage application started"); - ust_cmd.sock = -1; + ret = create_thread_poll_set(&events, 2); + if (ret < 0) { + goto error; + } + + ret = lttng_poll_add(&events, apps_cmd_pipe[0], LPOLLIN | LPOLLRDHUP); + if (ret < 0) { + goto error; + } while (1) { - /* See if we have a valid socket to add to pollfd */ - if (ust_cmd.sock != -1) { - nb_fd++; - update_poll_flag = 1; - } + /* Zeroed the events structure */ + lttng_poll_reset(&events); - /* The pollfd struct must be updated */ - if (update_poll_flag) { - ret = update_apps_cmd_pollfd(nb_fd, &pollfd); - if (ret < 0) { - /* malloc failed so we quit */ - goto error; - } - if (ust_cmd.sock != -1) { - /* Update pollfd with the new UST socket */ - DBG("Adding sock %d to apps cmd pollfd", ust_cmd.sock); - pollfd[nb_fd - 1].fd = ust_cmd.sock; - pollfd[nb_fd - 1].events = POLLHUP | POLLNVAL; - ust_cmd.sock = -1; - } - } + nb_fd = LTTNG_POLL_GETNB(&events); DBG("Apps thread polling on %d fds", nb_fd); /* Inifinite blocking call, waiting for transmission */ - ret = poll(pollfd, nb_fd, -1); + ret = lttng_poll_wait(&events, -1); if (ret < 0) { - perror("poll apps thread"); goto error; } - /* Thread quit pipe has been closed. Killing thread. */ - if (pollfd[0].revents == POLLNVAL) { - goto error; - } else if (pollfd[1].revents == POLLERR) { - ERR("Apps command pipe poll error"); - goto error; - } else if (pollfd[1].revents == POLLIN) { - /* Empty pipe */ - ret = read(apps_cmd_pipe[0], &ust_cmd, sizeof(ust_cmd)); - if (ret < 0 || ret < sizeof(ust_cmd)) { - perror("read apps cmd pipe"); - goto error; - } + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); - /* Register applicaton to the session daemon */ - ret = register_traceable_app(&ust_cmd.reg_msg, ust_cmd.sock); - if (ret < 0) { - /* Only critical ENOMEM error can be returned here */ + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { goto error; } - ret = send_ust_register_done(ust_cmd.sock); - if (ret < 0) { + /* Inspect the apps cmd pipe */ + if (pollfd == apps_cmd_pipe[0]) { + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + ERR("Apps command pipe error"); + goto error; + } else if (revents & LPOLLIN) { + /* Empty pipe */ + ret = read(apps_cmd_pipe[0], &ust_cmd, sizeof(ust_cmd)); + if (ret < 0 || ret < sizeof(ust_cmd)) { + perror("read apps cmd pipe"); + goto error; + } + + /* Register applicaton to the session daemon */ + ret = register_traceable_app(&ust_cmd.reg_msg, ust_cmd.sock); + if (ret < 0) { + /* Only critical ENOMEM error can be returned here */ + goto error; + } + + ret = ustctl_register_done(ust_cmd.sock); + if (ret < 0) { + /* + * If the registration is not possible, we simply + * unregister the apps and continue + */ + unregister_traceable_app(ust_cmd.sock); + } else { + /* + * We just need here to monitor the close of the UST + * socket and poll set monitor those by default. + */ + ret = lttng_poll_add(&events, ust_cmd.sock, 0); + if (ret < 0) { + goto error; + } + + DBG("Apps with sock %d added to poll set", ust_cmd.sock); + } + break; + } + } else { /* - * If the registration is not possible, we simply unregister - * the apps and continue + * At this point, we know that a registered application made the + * event at poll_wait. */ - unregister_traceable_app(ust_cmd.sock); - } - } + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + /* Removing from the poll set */ + ret = lttng_poll_del(&events, pollfd); + if (ret < 0) { + goto error; + } - count = nb_fd; - for (i = 2; i < count; i++) { - /* Apps socket is closed/hungup */ - switch (pollfd[i].revents) { - case POLLNVAL: - case POLLHUP: - /* Pipe closed */ - unregister_traceable_app(pollfd[i].fd); - nb_fd--; + /* Socket closed */ + unregister_traceable_app(pollfd); + break; + } } } - - if (nb_fd != count) { - update_poll_flag = 1; - } } error: @@ -971,7 +1022,7 @@ error: close(apps_cmd_pipe[0]); close(apps_cmd_pipe[1]); - free(pollfd); + lttng_poll_clean(&events); return NULL; } @@ -1003,8 +1054,12 @@ static void *thread_dispatch_ust_registration(void *data) ust_cmd = caa_container_of(node, struct ust_command, node); - DBG("Dispatching UST registration pid:%d sock:%d", - ust_cmd->reg_msg.pid, ust_cmd->sock); + DBG("Dispatching UST registration pid:%d ppid:%d uid:%d" + " gid:%d sock:%d name:%s (version %d.%d)", + ust_cmd->reg_msg.pid, ust_cmd->reg_msg.ppid, + ust_cmd->reg_msg.uid, ust_cmd->reg_msg.gid, + ust_cmd->sock, ust_cmd->reg_msg.name, + ust_cmd->reg_msg.major, ust_cmd->reg_msg.minor); /* * Inform apps thread of the new application registration. This * call is blocking so we can be assured that the data will be read @@ -1041,8 +1096,9 @@ error: */ static void *thread_registration_apps(void *data) { - int sock = 0, ret; - struct pollfd pollfd[2]; + int sock = 0, i, ret, pollfd; + uint32_t revents, nb_fd; + struct lttng_poll_event events; /* * Get allocated in this thread, enqueued to a global queue, dequeued and * freed in the manage apps thread. @@ -1056,87 +1112,122 @@ static void *thread_registration_apps(void *data) goto error; } - /* First fd is always the quit pipe */ - pollfd[0].fd = thread_quit_pipe[0]; + /* + * Pass 2 as size here for the thread quit pipe and apps socket. Nothing + * more will be added to this poll set. + */ + ret = create_thread_poll_set(&events, 2); + if (ret < 0) { + goto error; + } - /* Apps socket */ - pollfd[1].fd = apps_sock; - pollfd[1].events = POLLIN; + /* Add the application registration socket */ + ret = lttng_poll_add(&events, apps_sock, LPOLLIN | LPOLLRDHUP); + if (ret < 0) { + goto error; + } /* Notify all applications to register */ - //notify_apps(default_global_apps_pipe); + ret = notify_ust_apps(1); + if (ret < 0) { + ERR("Failed to notify applications or create the wait shared memory.\n" + "Execution continues but there might be problem for already running\n" + "applications that wishes to register."); + } while (1) { DBG("Accepting application registration"); + nb_fd = LTTNG_POLL_GETNB(&events); + /* Inifinite blocking call, waiting for transmission */ - ret = poll(pollfd, 2, -1); + ret = lttng_poll_wait(&events, -1); if (ret < 0) { - perror("poll register apps thread"); goto error; } - /* Thread quit pipe has been closed. Killing thread. */ - if (pollfd[0].revents == POLLNVAL) { - goto error; - } else if (pollfd[1].revents == POLLERR) { - ERR("Register apps socket poll error"); - goto error; - } + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); - sock = lttcomm_accept_unix_sock(apps_sock); - if (sock < 0) { - goto error; - } + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { + goto error; + } - /* Create UST registration command for enqueuing */ - ust_cmd = malloc(sizeof(struct ust_command)); - if (ust_cmd == NULL) { - perror("ust command malloc"); - goto error; - } + /* Event on the registration socket */ + if (pollfd == apps_sock) { + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + ERR("Register apps socket poll error"); + goto error; + } else if (revents & LPOLLIN) { + sock = lttcomm_accept_unix_sock(apps_sock); + if (sock < 0) { + goto error; + } - /* - * Using message-based transmissions to ensure we don't have to deal - * with partially received messages. - */ - ret = lttcomm_recv_unix_sock(sock, &ust_cmd->reg_msg, - sizeof(struct ust_register_msg)); - if (ret < 0 || ret != sizeof(struct ust_register_msg)) { - perror("lttcomm_recv_unix_sock register apps"); - free(ust_cmd); - close(sock); - continue; - } + /* Create UST registration command for enqueuing */ + ust_cmd = malloc(sizeof(struct ust_command)); + if (ust_cmd == NULL) { + perror("ust command malloc"); + goto error; + } - ust_cmd->sock = sock; + /* + * Using message-based transmissions to ensure we don't + * have to deal with partially received messages. + */ + ret = lttcomm_recv_unix_sock(sock, &ust_cmd->reg_msg, + sizeof(struct ust_register_msg)); + if (ret < 0 || ret < sizeof(struct ust_register_msg)) { + if (ret < 0) { + perror("lttcomm_recv_unix_sock register apps"); + } else { + ERR("Wrong size received on apps register"); + } + free(ust_cmd); + close(sock); + continue; + } - /* - * Lock free enqueue the registration request. - * The red pill has been taken! This apps will be part of the *system* - */ - cds_wfq_enqueue(&ust_cmd_queue.queue, &ust_cmd->node); + ust_cmd->sock = sock; - /* - * Wake the registration queue futex. - * Implicit memory barrier with the exchange in cds_wfq_enqueue. - */ - futex_nto1_wake(&ust_cmd_queue.futex); + DBG("UST registration received with pid:%d ppid:%d uid:%d" + " gid:%d sock:%d name:%s (version %d.%d)", + ust_cmd->reg_msg.pid, ust_cmd->reg_msg.ppid, + ust_cmd->reg_msg.uid, ust_cmd->reg_msg.gid, + ust_cmd->sock, ust_cmd->reg_msg.name, + ust_cmd->reg_msg.major, ust_cmd->reg_msg.minor); + /* + * Lock free enqueue the registration request. The red pill + * has been taken! This apps will be part of the *system* :) + */ + cds_wfq_enqueue(&ust_cmd_queue.queue, &ust_cmd->node); - DBG("Thread manage apps informed of queued node with sock:%d pid:%d", - sock, ust_cmd->reg_msg.pid); + /* + * Wake the registration queue futex. Implicit memory + * barrier with the exchange in cds_wfq_enqueue. + */ + futex_nto1_wake(&ust_cmd_queue.futex); + } + } + } } error: - DBG("Register apps thread dying"); - if (apps_sock) { - close(apps_sock); - } - if (sock) { - close(sock); - } + DBG("UST Registration thread dying"); + + /* Notify that the registration thread is gone */ + notify_ust_apps(0); + close(apps_sock); + close(sock); unlink(apps_unix_sock_path); + + lttng_poll_clean(&events); + return NULL; } @@ -1310,42 +1401,6 @@ error: return ret; } -/* - * modprobe_remove_kernel_modules - * Remove modules in reverse load order. - */ -static int modprobe_remove_kernel_modules(void) -{ - int ret = 0, i; - char modprobe[256]; - - for (i = ARRAY_SIZE(kernel_modules_list) - 1; i >= 0; i--) { - ret = snprintf(modprobe, sizeof(modprobe), - "/sbin/modprobe --remove --quiet %s", - kernel_modules_list[i].name); - if (ret < 0) { - perror("snprintf modprobe --remove"); - goto error; - } - modprobe[sizeof(modprobe) - 1] = '\0'; - ret = system(modprobe); - if (ret == -1) { - ERR("Unable to launch modprobe --remove for module %s", - kernel_modules_list[i].name); - } else if (kernel_modules_list[i].required - && WEXITSTATUS(ret) != 0) { - ERR("Unable to remove module %s", - kernel_modules_list[i].name); - } else { - DBG("Modprobe removal successful %s", - kernel_modules_list[i].name); - } - } - -error: - return ret; -} - /* * mount_debugfs */ @@ -1485,9 +1540,9 @@ error: } /* - * Notify kernel thread to update it's pollfd. + * Notify kernel thread to update it's poll set. */ -static int notify_kernel_pollfd(void) +static int notify_kernel_channels_update(void) { int ret; @@ -1541,6 +1596,43 @@ error: return NULL; } +/* + * Create an UST session and add it to the session ust list. + */ +static int create_ust_session(pid_t pid, struct ltt_session *session) +{ + int ret = -1; + struct ltt_ust_session *lus; + + DBG("Creating UST session"); + + lus = trace_ust_create_session(session->path, pid); + if (lus == NULL) { + goto error; + } + + ret = mkdir_recursive(lus->path, S_IRWXU | S_IRWXG, + geteuid(), allowed_group()); + if (ret < 0) { + if (ret != -EEXIST) { + ERR("Trace directory creation error"); + goto error; + } + } + + /* Create session on the UST tracer */ + ret = ustctl_create_session(lus); + if (ret < 0) { + goto error; + } + + return 0; + +error: + free(lus); + return ret; +} + /* * Create a kernel tracer session then create the default channel. */ @@ -1716,7 +1808,6 @@ static int process_client_msg(struct command_ctx *cmd_ctx) goto error; } } - /* Need a session for kernel command */ switch (cmd_ctx->lsm->cmd_type) { case LTTNG_CALIBRATE: @@ -1731,9 +1822,7 @@ static int process_client_msg(struct command_ctx *cmd_ctx) ret = LTTCOMM_KERN_SESS_FAIL; goto error; } - /* Start the kernel consumer daemon */ - if (kconsumerd_pid == 0 && cmd_ctx->lsm->cmd_type != LTTNG_REGISTER_CONSUMER) { ret = start_kconsumerd(); @@ -1744,6 +1833,8 @@ static int process_client_msg(struct command_ctx *cmd_ctx) } } break; + case LTTNG_DOMAIN_UST_PID: + break; default: break; } @@ -1922,11 +2013,13 @@ static int process_client_msg(struct command_ctx *cmd_ctx) switch (cmd_ctx->lsm->domain.type) { case LTTNG_DOMAIN_KERNEL: - kchan = trace_kernel_get_channel_by_name(cmd_ctx->lsm->u.enable.channel_name, + kchan = trace_kernel_get_channel_by_name( + cmd_ctx->lsm->u.enable.channel_name, cmd_ctx->session->kernel_session); if (kchan == NULL) { /* Channel not found, creating it */ - DBG("Creating kernel channel"); + DBG("Creating kernel channel %s", + cmd_ctx->lsm->u.enable.channel_name); ret = kernel_create_channel(cmd_ctx->session->kernel_session, &cmd_ctx->lsm->u.channel.chan, @@ -1937,7 +2030,7 @@ static int process_client_msg(struct command_ctx *cmd_ctx) } /* Notify kernel thread that there is a new channel */ - ret = notify_kernel_pollfd(); + ret = notify_kernel_channels_update(); if (ret < 0) { ret = LTTCOMM_FATAL; goto error; @@ -1953,9 +2046,11 @@ static int process_client_msg(struct command_ctx *cmd_ctx) } kernel_wait_quiescent(kernel_tracer_fd); + break; + case LTTNG_DOMAIN_UST_PID: + break; default: - /* TODO: Userspace tracing */ ret = LTTCOMM_NOT_IMPLEMENTED; goto error; } @@ -2005,6 +2100,12 @@ static int process_client_msg(struct command_ctx *cmd_ctx) ret = LTTCOMM_FATAL; goto error; } + + ret = notify_kernel_channels_update(); + if (ret < 0) { + ret = LTTCOMM_FATAL; + goto error; + } } kevent = trace_kernel_get_event_by_name(cmd_ctx->lsm->u.enable.event.name, kchan); @@ -2083,6 +2184,12 @@ static int process_client_msg(struct command_ctx *cmd_ctx) ret = LTTCOMM_FATAL; goto error; } + + ret = notify_kernel_channels_update(); + if (ret < 0) { + ret = LTTCOMM_FATAL; + goto error; + } } /* For each event in the kernel session */ @@ -2319,10 +2426,10 @@ static int process_client_msg(struct command_ctx *cmd_ctx) } /* - * Must notify the kernel thread here to update it's pollfd in order to - * remove the channel(s)' fd just destroyed. + * Must notify the kernel thread here to update it's poll setin order + * to remove the channel(s)' fd just destroyed. */ - ret = notify_kernel_pollfd(); + ret = notify_kernel_channels_update(); if (ret < 0) { ret = LTTCOMM_FATAL; goto error; @@ -2339,7 +2446,7 @@ static int process_client_msg(struct command_ctx *cmd_ctx) nb_dom++; } - nb_dom += cmd_ctx->session->ust_trace_count; + nb_dom += cmd_ctx->session->ust_session_list.count; ret = setup_lttng_msg(cmd_ctx, sizeof(struct lttng_domain) * nb_dom); if (ret < 0) { @@ -2544,9 +2651,10 @@ setup_error: */ static void *thread_manage_clients(void *data) { - int sock = 0, ret; + int sock = 0, ret, i, pollfd; + uint32_t revents, nb_fd; struct command_ctx *cmd_ctx = NULL; - struct pollfd pollfd[2]; + struct lttng_poll_event events; DBG("[thread] Manage client started"); @@ -2555,15 +2663,23 @@ static void *thread_manage_clients(void *data) goto error; } - /* First fd is always the quit pipe */ - pollfd[0].fd = thread_quit_pipe[0]; + /* + * Pass 2 as size here for the thread quit pipe and client_sock. Nothing + * more will be added to this poll set. + */ + ret = create_thread_poll_set(&events, 2); + if (ret < 0) { + goto error; + } - /* Apps socket */ - pollfd[1].fd = client_sock; - pollfd[1].events = POLLIN; + /* Add the application registration socket */ + ret = lttng_poll_add(&events, client_sock, LPOLLIN | LPOLLPRI); + if (ret < 0) { + goto error; + } - /* Notify parent pid that we are ready - * to accept command for client side. + /* + * Notify parent pid that we are ready to accept command for client side. */ if (opt_sig_parent) { kill(ppid, SIGCHLD); @@ -2572,21 +2688,36 @@ static void *thread_manage_clients(void *data) while (1) { DBG("Accepting client command ..."); + nb_fd = LTTNG_POLL_GETNB(&events); + /* Inifinite blocking call, waiting for transmission */ - ret = poll(pollfd, 2, -1); + ret = lttng_poll_wait(&events, -1); if (ret < 0) { - perror("poll client thread"); goto error; } - /* Thread quit pipe has been closed. Killing thread. */ - if (pollfd[0].revents == POLLNVAL) { - goto error; - } else if (pollfd[1].revents == POLLERR) { - ERR("Client socket poll error"); - goto error; + for (i = 0; i < nb_fd; i++) { + /* Fetch once the poll data */ + revents = LTTNG_POLL_GETEV(&events, i); + pollfd = LTTNG_POLL_GETFD(&events, i); + + /* Thread quit pipe has been closed. Killing thread. */ + ret = check_thread_quit_pipe(pollfd, revents); + if (ret) { + goto error; + } + + /* Event on the registration socket */ + if (pollfd == client_sock) { + if (revents & (LPOLLERR | LPOLLHUP | LPOLLRDHUP)) { + ERR("Client socket poll error"); + goto error; + } + } } + DBG("Wait for client response"); + sock = lttcomm_accept_unix_sock(client_sock); if (sock < 0) { goto error; @@ -2594,9 +2725,18 @@ static void *thread_manage_clients(void *data) /* Allocate context command to process the client request */ cmd_ctx = malloc(sizeof(struct command_ctx)); + if (cmd_ctx == NULL) { + perror("malloc cmd_ctx"); + goto error; + } /* Allocate data buffer for reception */ cmd_ctx->lsm = malloc(sizeof(struct lttcomm_session_msg)); + if (cmd_ctx->lsm == NULL) { + perror("malloc cmd_ctx->lsm"); + goto error; + } + cmd_ctx->llm = NULL; cmd_ctx->session = NULL; @@ -2606,12 +2746,17 @@ static void *thread_manage_clients(void *data) * the client. */ DBG("Receiving data from client ..."); - ret = lttcomm_recv_unix_sock(sock, cmd_ctx->lsm, sizeof(struct lttcomm_session_msg)); + ret = lttcomm_recv_unix_sock(sock, cmd_ctx->lsm, + sizeof(struct lttcomm_session_msg)); if (ret <= 0) { + DBG("Nothing recv() from client... continuing"); + close(sock); + free(cmd_ctx); continue; } - // TODO: Validate cmd_ctx including sanity check for security purpose. + // TODO: Validate cmd_ctx including sanity check for + // security purpose. /* * This function dispatch the work to the kernel or userspace tracer @@ -2621,16 +2766,19 @@ static void *thread_manage_clients(void *data) */ ret = process_client_msg(cmd_ctx); if (ret < 0) { - /* TODO: Inform client somehow of the fatal error. At this point, - * ret < 0 means that a malloc failed (ENOMEM). */ - /* Error detected but still accept command */ + /* + * TODO: Inform client somehow of the fatal error. At + * this point, ret < 0 means that a malloc failed + * (ENOMEM). Error detected but still accept command. + */ clean_command_ctx(&cmd_ctx); continue; } DBG("Sending response (size: %d, retcode: %d)", cmd_ctx->lttng_msg_size, cmd_ctx->llm->ret_code); - ret = send_unix_sock(sock, cmd_ctx->llm, cmd_ctx->lttng_msg_size); + ret = send_unix_sock(sock, cmd_ctx->llm, + cmd_ctx->lttng_msg_size); if (ret < 0) { ERR("Failed to send data back to client"); } @@ -2643,15 +2791,11 @@ static void *thread_manage_clients(void *data) error: DBG("Client thread dying"); - if (client_sock) { - close(client_sock); - } - if (sock) { - close(sock); - } - unlink(client_unix_sock_path); + close(client_sock); + close(sock); + lttng_poll_clean(&events); clean_command_ctx(&cmd_ctx); return NULL; } @@ -3089,6 +3233,12 @@ int main(int argc, char **argv) snprintf(client_unix_sock_path, PATH_MAX, DEFAULT_GLOBAL_CLIENT_UNIX_SOCK); } + + /* Set global SHM for ust */ + if (strlen(wait_shm_path) == 0) { + snprintf(wait_shm_path, PATH_MAX, + DEFAULT_GLOBAL_APPS_WAIT_SHM_PATH); + } } else { home_path = get_home_dir(); if (home_path == NULL) { @@ -3108,6 +3258,12 @@ int main(int argc, char **argv) snprintf(client_unix_sock_path, PATH_MAX, DEFAULT_HOME_CLIENT_UNIX_SOCK, home_path); } + + /* Set global SHM for ust */ + if (strlen(wait_shm_path) == 0) { + snprintf(wait_shm_path, PATH_MAX, + DEFAULT_HOME_APPS_WAIT_SHM_PATH, geteuid()); + } } DBG("Client socket path %s", client_unix_sock_path); @@ -3184,6 +3340,9 @@ int main(int argc, char **argv) */ session_list_ptr = get_session_list(); + /* Set up max poll set size */ + lttng_poll_set_max_size(); + /* Create thread to manage the client socket */ ret = pthread_create(&client_thread, NULL, thread_manage_clients, (void *) NULL);