#include "lttng-filter.h"
#include <lttng/align.h>
+#include "ust-events-internal.h"
static int lttng_fls(int val)
{
break;
case OBJECT_TYPE_U8:
dbg_printf("op load field u8\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
insn->op = FILTER_OP_LOAD_FIELD_U8;
break;
case OBJECT_TYPE_U16:
dbg_printf("op load field u16\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U16;
break;
case OBJECT_TYPE_U32:
dbg_printf("op load field u32\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U32;
break;
case OBJECT_TYPE_U64:
dbg_printf("op load field u64\n");
- stack_top->type = REG_S64;
+ stack_top->type = REG_U64;
if (!stack_top->load.rev_bo)
insn->op = FILTER_OP_LOAD_FIELD_U64;
break;
struct vstack_load *load, bool is_context)
{
load->type = LOAD_OBJECT;
- /*
- * LTTng-UST layout all integer fields as s64 on the stack for the filter.
- */
+
switch (field->type.atype) {
case atype_integer:
if (field->type.u.integer.signedness)
memset(&gid, 0, sizeof(gid));
gid.ctx_index = idx;
gid.elem.type = load->object_type;
+ gid.field = field;
data_offset = bytecode_push_data(runtime, &gid,
__alignof__(gid), sizeof(gid));
if (data_offset < 0) {
memset(&gid, 0, sizeof(gid));
gid.ctx_index = idx;
gid.elem.type = load->object_type;
+ gid.field = field;
data_offset = bytecode_push_data(runtime, &gid,
__alignof__(gid), sizeof(gid));
if (data_offset < 0) {
return ret;
}
-static int specialize_event_payload_lookup(struct lttng_event *event,
+static int specialize_payload_lookup(const struct lttng_event_desc *event_desc,
struct bytecode_runtime *runtime,
struct load_op *insn,
struct vstack_load *load)
{
const char *name;
uint16_t offset;
- const struct lttng_event_desc *desc = event->desc;
unsigned int i, nr_fields;
bool found = false;
uint32_t field_offset = 0;
struct filter_get_index_data gid;
ssize_t data_offset;
- nr_fields = desc->nr_fields;
+ nr_fields = event_desc->nr_fields;
offset = ((struct get_symbol *) insn->data)->offset;
name = runtime->p.bc->bc.data + runtime->p.bc->bc.reloc_offset + offset;
for (i = 0; i < nr_fields; i++) {
- field = &desc->fields[i];
+ field = &event_desc->fields[i];
if (field->u.ext.nofilter) {
continue;
}
memset(&gid, 0, sizeof(gid));
gid.offset = field_offset;
gid.elem.type = load->object_type;
+ gid.field = field;
data_offset = bytecode_push_data(runtime, &gid,
__alignof__(gid), sizeof(gid));
if (data_offset < 0) {
return ret;
}
-int lttng_filter_specialize_bytecode(struct lttng_event *event,
+int lttng_filter_specialize_bytecode(const struct lttng_event_desc *event_desc,
struct bytecode_runtime *bytecode)
{
void *pc, *next_pc, *start_pc;
goto end;
case FILTER_OP_RETURN:
- if (vstack_ax(stack)->type == REG_S64)
+ if (vstack_ax(stack)->type == REG_S64 ||
+ vstack_ax(stack)->type == REG_U64)
*(filter_opcode_t *) pc = FILTER_OP_RETURN_S64;
ret = 0;
goto end;
case FILTER_OP_RETURN_S64:
- if (vstack_ax(stack)->type != REG_S64) {
+ if (vstack_ax(stack)->type != REG_S64 &&
+ vstack_ax(stack)->type != REG_U64) {
ERR("Unexpected register type\n");
ret = -EINVAL;
goto end;
insn->op = FILTER_OP_EQ_STAR_GLOB_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_EQ_S64;
else
insn->op = FILTER_OP_EQ_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_EQ_S64_DOUBLE;
else
insn->op = FILTER_OP_EQ_DOUBLE;
insn->op = FILTER_OP_NE_STAR_GLOB_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_NE_S64;
else
insn->op = FILTER_OP_NE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_NE_S64_DOUBLE;
else
insn->op = FILTER_OP_NE_DOUBLE;
insn->op = FILTER_OP_GT_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GT_S64;
else
insn->op = FILTER_OP_GT_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GT_S64_DOUBLE;
else
insn->op = FILTER_OP_GT_DOUBLE;
insn->op = FILTER_OP_LT_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LT_S64;
else
insn->op = FILTER_OP_LT_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LT_S64_DOUBLE;
else
insn->op = FILTER_OP_LT_DOUBLE;
insn->op = FILTER_OP_GE_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GE_S64;
else
insn->op = FILTER_OP_GE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_GE_S64_DOUBLE;
else
insn->op = FILTER_OP_GE_DOUBLE;
ret = -EINVAL;
goto end;
}
- vstack_ax(stack)->type = REG_S64;
+ vstack_ax(stack)->type = REG_U64;
next_pc += sizeof(struct binary_op);
break;
}
insn->op = FILTER_OP_LE_STRING;
break;
case REG_S64:
+ case REG_U64:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LE_S64;
else
insn->op = FILTER_OP_LE_DOUBLE_S64;
case REG_DOUBLE:
if (vstack_bx(stack)->type == REG_UNKNOWN)
break;
- if (vstack_bx(stack)->type == REG_S64)
+ if (vstack_bx(stack)->type == REG_S64 ||
+ vstack_bx(stack)->type == REG_U64)
insn->op = FILTER_OP_LE_S64_DOUBLE;
else
insn->op = FILTER_OP_LE_DOUBLE;
case FILTER_OP_LT_S64_DOUBLE:
case FILTER_OP_GE_S64_DOUBLE:
case FILTER_OP_LE_S64_DOUBLE:
+ {
+ /* Pop 2, push 1 */
+ if (vstack_pop(stack)) {
+ ret = -EINVAL;
+ goto end;
+ }
+ vstack_ax(stack)->type = REG_S64;
+ next_pc += sizeof(struct binary_op);
+ break;
+ }
+
case FILTER_OP_BIT_RSHIFT:
case FILTER_OP_BIT_LSHIFT:
case FILTER_OP_BIT_AND:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_PLUS_S64;
break;
case REG_DOUBLE:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_MINUS_S64;
break;
case REG_DOUBLE:
goto end;
case REG_S64:
+ case REG_U64:
insn->op = FILTER_OP_UNARY_NOT_S64;
break;
case REG_DOUBLE:
insn->op = FILTER_OP_CAST_DOUBLE_TO_S64;
break;
case REG_UNKNOWN:
+ case REG_U64:
break;
}
/* Pop 1, push 1 */
case FILTER_OP_LOAD_FIELD_S16:
case FILTER_OP_LOAD_FIELD_S32:
case FILTER_OP_LOAD_FIELD_S64:
+ {
+ /* Pop 1, push 1 */
+ vstack_ax(stack)->type = REG_S64;
+ next_pc += sizeof(struct load_op);
+ break;
+ }
+
case FILTER_OP_LOAD_FIELD_U8:
case FILTER_OP_LOAD_FIELD_U16:
case FILTER_OP_LOAD_FIELD_U32:
case FILTER_OP_LOAD_FIELD_U64:
{
/* Pop 1, push 1 */
- vstack_ax(stack)->type = REG_S64;
+ vstack_ax(stack)->type = REG_U64;
next_pc += sizeof(struct load_op);
break;
}
break;
case LOAD_ROOT_PAYLOAD:
/* Lookup event payload field. */
- ret = specialize_event_payload_lookup(event,
+ ret = specialize_payload_lookup(event_desc,
bytecode, insn,
&vstack_ax(stack)->load);
if (ret)