/*******************************************************************************
- * Copyright (c) 2010, 2015 Ericsson, École Polytechnique de Montréal, and others
+ * Copyright (c) 2010, 2016 Ericsson, École Polytechnique de Montréal, and others
*
* All rights reserved. This program and the accompanying materials are
* made available under the terms of the Eclipse Public License v1.0 which
import java.util.List;
import java.util.concurrent.locks.ReentrantReadWriteLock;
+import org.eclipse.jdt.annotation.NonNull;
import org.eclipse.tracecompass.statesystem.core.exceptions.TimeRangeException;
import org.eclipse.tracecompass.statesystem.core.interval.ITmfStateInterval;
import org.eclipse.tracecompass.statesystem.core.statevalue.TmfStateValue;
+import com.google.common.collect.Iterables;
+
/**
* The base class for all the types of nodes that go in the History Tree.
*
* 1 - byte (done or not)
* </pre>
*/
- private static final int COMMON_HEADER_SIZE = 34;
+ private static final int COMMON_HEADER_SIZE = Byte.BYTES
+ + 2 * Long.BYTES
+ + 4 * Integer.BYTES
+ + Byte.BYTES;
// ------------------------------------------------------------------------
// Attributes
private final int fSequenceNumber;
private int fParentSequenceNumber; /* = -1 if this node is the root node */
- /* Where the Strings section begins (from the start of the node */
- private int fStringSectionOffset;
-
/* Sum of bytes of all intervals in the node */
private int fSizeOfIntervalSection;
fSequenceNumber = seqNumber;
fParentSequenceNumber = parentSeqNumber;
- fStringSectionOffset = config.getBlockSize();
fSizeOfIntervalSection = 0;
fIsOnDisk = false;
fIntervals = new ArrayList<>();
* @param fc
* FileChannel to the history file, ALREADY SEEKED at the start
* of the node.
+ * @param nodeFactory
+ * The factory to create the nodes for this tree
* @return The node object
* @throws IOException
* If there was an error reading from the file channel
*/
- public static final HTNode readNode(HTConfig config, FileChannel fc)
+ public static final @NonNull HTNode readNode(HTConfig config, FileChannel fc, IHistoryTree.IHTNodeFactory nodeFactory)
throws IOException {
HTNode newNode = null;
int res, i;
int seqNb = buffer.getInt();
int parentSeqNb = buffer.getInt();
int intervalCount = buffer.getInt();
- int stringSectionOffset = buffer.getInt();
buffer.get(); // TODO Used to be "isDone", to be removed from the header
/* Now the rest of the header depends on the node type */
switch (type) {
case CORE:
/* Core nodes */
- newNode = new CoreNode(config, seqNb, parentSeqNb, start);
+ newNode = nodeFactory.createCoreNode(config, seqNb, parentSeqNb, start);
newNode.readSpecificHeader(buffer);
break;
case LEAF:
/* Leaf nodes */
- newNode = new LeafNode(config, seqNb, parentSeqNb, start);
+ newNode = nodeFactory.createLeafNode(config, seqNb, parentSeqNb, start);
newNode.readSpecificHeader(buffer);
break;
for (i = 0; i < intervalCount; i++) {
HTInterval interval = HTInterval.readFrom(buffer);
newNode.fIntervals.add(interval);
- newNode.fSizeOfIntervalSection += interval.getIntervalSize();
+ newNode.fSizeOfIntervalSection += interval.getSizeOnDisk();
}
/* Assign the node's other information we have read previously */
newNode.fNodeEnd = end;
- newNode.fStringSectionOffset = stringSectionOffset;
newNode.fIsOnDisk = true;
return newNode;
fRwl.readLock().lock();
try {
final int blockSize = fConfig.getBlockSize();
- int curStringsEntryEndPos = blockSize;
ByteBuffer buffer = ByteBuffer.allocate(blockSize);
buffer.order(ByteOrder.LITTLE_ENDIAN);
buffer.putInt(fSequenceNumber);
buffer.putInt(fParentSequenceNumber);
buffer.putInt(fIntervals.size());
- buffer.putInt(fStringSectionOffset);
buffer.put((byte) 1); // TODO Used to be "isDone", to be removed from header
/* Now call the inner method to write the specific header part */
writeSpecificHeader(buffer);
/* Back to us, we write the intervals */
- for (HTInterval interval : fIntervals) {
- int size = interval.writeInterval(buffer, curStringsEntryEndPos);
- curStringsEntryEndPos -= size;
- }
+ fIntervals.forEach(i -> i.writeInterval(buffer));
/*
- * Write padding between the end of the Data section and the start
- * of the Strings section (needed to fill the node in case there is
- * no Strings section)
+ * Fill the rest with zeros
*/
- while (buffer.position() < fStringSectionOffset) {
+ while (buffer.position() < blockSize) {
buffer.put((byte) 0);
}
- /*
- * If the offsets were right, the size of the Strings section should
- * be == to the expected size
- */
- assert (curStringsEntryEndPos == fStringSectionOffset);
-
/* Finally, write everything in the Buffer to disk */
-
- // if we don't do this, flip() will lose what's after.
- buffer.position(blockSize);
-
buffer.flip();
int res = fc.write(buffer);
- assert (res == blockSize);
+ if (res != blockSize) {
+ throw new IllegalStateException("Wrong size of block written: Actual: " + res + ", Expected: " + blockSize); //$NON-NLS-1$ //$NON-NLS-2$
+ }
} finally {
fRwl.readLock().unlock();
fRwl.writeLock().lock();
try {
/* Just in case, should be checked before even calling this function */
- assert (newInterval.getIntervalSize() <= getNodeFreeSpace());
+ assert (newInterval.getSizeOnDisk() <= getNodeFreeSpace());
/* Find the insert position to keep the list sorted */
int index = fIntervals.size();
}
fIntervals.add(index, newInterval);
- fSizeOfIntervalSection += newInterval.getIntervalSize();
+ fSizeOfIntervalSection += newInterval.getSizeOnDisk();
- /* Update the in-node offset "pointer" */
- fStringSectionOffset -= (newInterval.getStringsEntrySize());
} finally {
fRwl.writeLock().unlock();
}
public void closeThisNode(long endtime) {
fRwl.writeLock().lock();
try {
- assert (endtime >= fNodeStart);
+ /**
+ * FIXME: was assert (endtime >= fNodeStart); but that exception
+ * is reached with an empty node that has start time endtime + 1
+ */
+// if (endtime < fNodeStart) {
+// throw new IllegalArgumentException("Endtime " + endtime + " cannot be lower than start time " + fNodeStart);
+// }
if (!fIntervals.isEmpty()) {
/*
* EndTime > the one requested. Only need to check the last one
* since they are sorted
*/
- assert (endtime >= fIntervals.get(fIntervals.size() - 1).getEndTime());
+ if (endtime < Iterables.getLast(fIntervals).getEndTime()) {
+ throw new IllegalArgumentException("Closing end time should be greater than or equal to the end time of the intervals of this node"); //$NON-NLS-1$
+ }
}
fNodeEnd = endtime;
* null anyway).
*/
ITmfStateInterval interval = fIntervals.get(i);
- if (interval.getStartTime() <= t &&
+ if (t >= interval.getStartTime() &&
interval.getAttribute() < stateInfo.size()) {
stateInfo.set(interval.getAttribute(), interval);
}
*/
index = -index - 1;
- }
-
- /* Sometimes binarySearch yields weird stuff... */
- if (index < 0) {
- index = 0;
- }
- if (index >= fIntervals.size()) {
- index = fIntervals.size() - 1;
- }
-
- /*
- * Another API quirkiness, the returned index is the one of the *last*
- * element of a series of equal endtimes, which happens sometimes. We
- * want the *first* element of such a series, to read through them
- * again.
- */
- while (index > 0
- && fIntervals.get(index - 1).compareTo(fIntervals.get(index)) == 0) {
- index--;
+ } else {
+ /*
+ * Another API quirkiness, the returned index is the one of the *last*
+ * element of a series of equal endtimes, which happens sometimes. We
+ * want the *first* element of such a series, to read through them
+ * again.
+ */
+ while (index > 0
+ && fIntervals.get(index - 1).compareTo(fIntervals.get(index)) == 0) {
+ index--;
+ }
}
return index;
*/
public int getNodeFreeSpace() {
fRwl.readLock().lock();
- int ret = fStringSectionOffset - getDataSectionEndOffset();
+ int ret = fConfig.getBlockSize() - getDataSectionEndOffset();
fRwl.readLock().unlock();
return ret;
@Override
public String toString() {
/* Only used for debugging, shouldn't be externalized */
- StringBuffer buf = new StringBuffer("Node #" + fSequenceNumber + ", ");
- buf.append(toStringSpecific());
- buf.append(fIntervals.size() + " intervals (" + getNodeUsagePercent()
- + "% used), ");
-
- buf.append("[" + fNodeStart + " - ");
- if (fIsOnDisk) {
- buf = buf.append("" + fNodeEnd + "]");
- } else {
- buf = buf.append("...]");
- }
- return buf.toString();
+ return String.format("Node #%d, %s, %s, %d intervals (%d%% used), [%d - %s]",
+ fSequenceNumber,
+ (fParentSequenceNumber == -1) ? "Root" : "Parent #" + fParentSequenceNumber,
+ toStringSpecific(),
+ fIntervals.size(),
+ getNodeUsagePercent(),
+ fNodeStart,
+ (fIsOnDisk || fNodeEnd != 0) ? fNodeEnd : "...");
}
/**
@SuppressWarnings("nls")
public void debugPrintIntervals(PrintWriter writer) {
/* Only used for debugging, shouldn't be externalized */
- writer.println("Node #" + fSequenceNumber + ":");
+ writer.println("Intervals for node #" + fSequenceNumber + ":");
/* Array of children */
- if (getNodeType() == NodeType.CORE) { /* Only Core Nodes can have children */
- CoreNode thisNode = (CoreNode) this;
+ if (getNodeType() != NodeType.LEAF) { /* Only Core Nodes can have children */
+ ParentNode thisNode = (ParentNode) this;
writer.print(" " + thisNode.getNbChildren() + " children");
if (thisNode.getNbChildren() >= 1) {
writer.print(": [ " + thisNode.getChild(0));