Fix data_offset when importing the indexes
[babeltrace.git] / formats / ctf / ctf.c
index 37f339c92a107b00695052bc238634ad1fe7224f..eec3350f3ba73b2c40c27c479943f3d80b4c093c 100644 (file)
@@ -471,6 +471,15 @@ int ctf_read_event(struct bt_stream_pos *ppos, struct ctf_stream_definition *str
         */
        if (unlikely(pos->offset == EOF))
                return EOF;
+
+       if (pos->content_size == 0) {
+               /* Stream is inactive for now (live reading). */
+               return EAGAIN;
+       }
+       /* Packet only contains headers */
+       if (pos->offset == pos->content_size)
+               return EAGAIN;
+
        assert(pos->offset < pos->content_size);
 
        /* Read event header */
@@ -618,6 +627,107 @@ error:
        return ret;
 }
 
+/*
+ * One side-effect of this function is to unmap pos mmap base if one is
+ * mapped.
+ */
+static
+int find_data_offset(struct ctf_stream_pos *pos,
+               struct ctf_file_stream *file_stream,
+               struct packet_index *packet_index)
+{
+       uint64_t packet_map_len = DEFAULT_HEADER_LEN, tmp_map_len;
+       struct stat filestats;
+       size_t filesize;
+       int ret;
+
+       pos = &file_stream->pos;
+
+       ret = fstat(pos->fd, &filestats);
+       if (ret < 0)
+               return ret;
+       filesize = filestats.st_size;
+
+       /* Deal with empty files */
+       if (!filesize) {
+               return 0;
+       }
+
+begin:
+       if (filesize - pos->mmap_offset < (packet_map_len >> LOG2_CHAR_BIT)) {
+               packet_map_len = (filesize - pos->mmap_offset) << LOG2_CHAR_BIT;
+       }
+
+       if (pos->base_mma) {
+               /* unmap old base */
+               ret = munmap_align(pos->base_mma);
+               if (ret) {
+                       fprintf(stderr, "[error] Unable to unmap old base: %s.\n",
+                                       strerror(errno));
+                       return ret;
+               }
+               pos->base_mma = NULL;
+       }
+       /* map new base. Need mapping length from header. */
+       pos->base_mma = mmap_align(packet_map_len >> LOG2_CHAR_BIT, PROT_READ,
+                       MAP_PRIVATE, pos->fd, pos->mmap_offset);
+       assert(pos->base_mma != MAP_FAILED);
+
+       pos->content_size = packet_map_len;
+       pos->packet_size = packet_map_len;
+       pos->offset = 0;        /* Position of the packet header */
+
+       /* update trace_packet_header and stream_packet_context */
+       if (pos->prot == PROT_READ && file_stream->parent.trace_packet_header) {
+               /* Read packet header */
+               ret = generic_rw(&pos->parent, &file_stream->parent.trace_packet_header->p);
+               if (ret) {
+                       if (ret == -EFAULT)
+                               goto retry;
+               }
+       }
+       if (pos->prot == PROT_READ && file_stream->parent.stream_packet_context) {
+               /* Read packet context */
+               ret = generic_rw(&pos->parent, &file_stream->parent.stream_packet_context->p);
+               if (ret) {
+                       if (ret == -EFAULT)
+                               goto retry;
+               }
+       }
+       packet_index->data_offset = pos->offset;
+
+       /* unmap old base */
+       ret = munmap_align(pos->base_mma);
+       if (ret) {
+               fprintf(stderr, "[error] Unable to unmap old base: %s.\n",
+                               strerror(errno));
+               return ret;
+       }
+       pos->base_mma = NULL;
+
+       return 0;
+
+       /* Retry with larger mapping */
+retry:
+       if (packet_map_len == ((filesize - pos->mmap_offset) << LOG2_CHAR_BIT)) {
+               /*
+                * Reached EOF, but still expecting header/context data.
+                */
+               fprintf(stderr, "[error] Reached end of file, but still expecting header or context fields.\n");
+               return -EFAULT;
+       }
+       /* Double the mapping len, and retry */
+       tmp_map_len = packet_map_len << 1;
+       if (tmp_map_len >> 1 != packet_map_len) {
+               /* Overflow */
+               fprintf(stderr, "[error] Packet mapping length overflow\n");
+               return -EFAULT;
+       }
+       packet_map_len = tmp_map_len;
+       goto begin;
+}
+
+
 int ctf_init_pos(struct ctf_stream_pos *pos, struct bt_trace_descriptor *trace,
                int fd, int open_flags)
 {
@@ -836,11 +946,17 @@ void ctf_packet_seek(struct bt_stream_pos *stream_pos, size_t index, int whence)
                                struct packet_index,
                                pos->cur_index);
                file_stream->parent.real_timestamp = packet_index->timestamp_begin;
-               pos->mmap_offset = packet_index->offset;
 
                /* Lookup context/packet size in index */
+               if (packet_index->data_offset == -1) {
+                       ret = find_data_offset(pos, file_stream, packet_index);
+                       if (ret < 0) {
+                               return;
+                       }
+               }
                pos->content_size = packet_index->content_size;
                pos->packet_size = packet_index->packet_size;
+               pos->mmap_offset = packet_index->offset;
                if (packet_index->data_offset < packet_index->content_size) {
                        pos->offset = 0;        /* will read headers */
                } else if (packet_index->data_offset == packet_index->content_size) {
@@ -1554,9 +1670,9 @@ begin:
                }
        } else {
                /* Use file size for packet size */
-               packet_index.content_size = filesize * CHAR_BIT;
-               /* Use content size if non-zero, else file size */
-               packet_index.packet_size = packet_index.content_size ? : filesize * CHAR_BIT;
+               packet_index.packet_size = filesize * CHAR_BIT;
+               /* Use packet size if non-zero, else file size */
+               packet_index.content_size = packet_index.packet_size ? : filesize * CHAR_BIT;
        }
 
        /* Validate content size and packet size values */
@@ -1572,6 +1688,16 @@ begin:
                return -EINVAL;
        }
 
+       if (packet_index.content_size < pos->offset) {
+               fprintf(stderr, "[error] Invalid CTF stream: content size is smaller than packet headers.\n");
+               return -EINVAL;
+       }
+
+       if ((packet_index.packet_size >> LOG2_CHAR_BIT) == 0) {
+               fprintf(stderr, "[error] Invalid CTF stream: packet size needs to be at least one byte\n");
+               return -EINVAL;
+       }
+
        /* Save position after header and context */
        packet_index.data_offset = pos->offset;
 
@@ -1711,6 +1837,11 @@ int import_stream_packet_index(struct ctf_trace *td,
                ret = -1;
                goto error;
        }
+       if (index_hdr.packet_index_len == 0) {
+               fprintf(stderr, "[error] Packet index length cannot be 0.\n");
+               ret = -1;
+               goto error;
+       }
 
        while ((index_read = fread(&ctf_index, index_hdr.packet_index_len, 1,
                                        pos->index_fp)) == 1) {
@@ -1724,6 +1855,7 @@ int import_stream_packet_index(struct ctf_trace *td,
                index.timestamp_end = be64toh(ctf_index.timestamp_end);
                index.events_discarded = be64toh(ctf_index.events_discarded);
                index.events_discarded_len = 64;
+               index.data_offset = -1;
                stream_id = be64toh(ctf_index.stream_id);
 
                if (!first_packet) {
@@ -1844,6 +1976,10 @@ int ctf_open_file_stream_read(struct ctf_trace *td, const char *path, int flags,
                        goto error_free;
                }
                file_stream->pos.index_fp = fdopen(ret, "r");
+               if (!file_stream->pos.index_fp) {
+                       perror("fdopen() error");
+                       goto error_free;
+               }
                ret = import_stream_packet_index(td, file_stream);
                if (ret) {
                        ret = -1;
This page took 0.025349 seconds and 4 git commands to generate.