7 #include "thread-stack.h"
13 int thread__init_map_groups(struct thread
*thread
, struct machine
*machine
)
15 struct thread
*leader
;
16 pid_t pid
= thread
->pid_
;
18 if (pid
== thread
->tid
|| pid
== -1) {
19 thread
->mg
= map_groups__new(machine
);
21 leader
= __machine__findnew_thread(machine
, pid
, pid
);
23 thread
->mg
= map_groups__get(leader
->mg
);
26 return thread
->mg
? 0 : -1;
29 struct thread
*thread__new(pid_t pid
, pid_t tid
)
33 struct thread
*thread
= zalloc(sizeof(*thread
));
40 INIT_LIST_HEAD(&thread
->comm_list
);
42 if (unwind__prepare_access(thread
) < 0)
45 comm_str
= malloc(32);
49 snprintf(comm_str
, 32, ":%d", tid
);
50 comm
= comm__new(comm_str
, 0, false);
55 list_add(&comm
->list
, &thread
->comm_list
);
56 atomic_set(&thread
->refcnt
, 0);
57 RB_CLEAR_NODE(&thread
->rb_node
);
67 void thread__delete(struct thread
*thread
)
69 struct comm
*comm
, *tmp
;
71 BUG_ON(!RB_EMPTY_NODE(&thread
->rb_node
));
73 thread_stack__free(thread
);
76 map_groups__put(thread
->mg
);
79 list_for_each_entry_safe(comm
, tmp
, &thread
->comm_list
, list
) {
80 list_del(&comm
->list
);
83 unwind__finish_access(thread
);
88 struct thread
*thread__get(struct thread
*thread
)
91 atomic_inc(&thread
->refcnt
);
95 void thread__put(struct thread
*thread
)
97 if (thread
&& atomic_dec_and_test(&thread
->refcnt
)) {
98 list_del_init(&thread
->node
);
99 thread__delete(thread
);
103 struct comm
*thread__comm(const struct thread
*thread
)
105 if (list_empty(&thread
->comm_list
))
108 return list_first_entry(&thread
->comm_list
, struct comm
, list
);
111 struct comm
*thread__exec_comm(const struct thread
*thread
)
113 struct comm
*comm
, *last
= NULL
;
115 list_for_each_entry(comm
, &thread
->comm_list
, list
) {
124 int __thread__set_comm(struct thread
*thread
, const char *str
, u64 timestamp
,
127 struct comm
*new, *curr
= thread__comm(thread
);
130 /* Override the default :tid entry */
131 if (!thread
->comm_set
) {
132 err
= comm__override(curr
, str
, timestamp
, exec
);
136 new = comm__new(str
, timestamp
, exec
);
139 list_add(&new->list
, &thread
->comm_list
);
142 unwind__flush_access(thread
);
145 thread
->comm_set
= true;
150 const char *thread__comm_str(const struct thread
*thread
)
152 const struct comm
*comm
= thread__comm(thread
);
157 return comm__str(comm
);
160 /* CHECKME: it should probably better return the max comm len from its comm list */
161 int thread__comm_len(struct thread
*thread
)
163 if (!thread
->comm_len
) {
164 const char *comm
= thread__comm_str(thread
);
167 thread
->comm_len
= strlen(comm
);
170 return thread
->comm_len
;
173 size_t thread__fprintf(struct thread
*thread
, FILE *fp
)
175 return fprintf(fp
, "Thread %d %s\n", thread
->tid
, thread__comm_str(thread
)) +
176 map_groups__fprintf(thread
->mg
, fp
);
179 void thread__insert_map(struct thread
*thread
, struct map
*map
)
181 map_groups__fixup_overlappings(thread
->mg
, map
, stderr
);
182 map_groups__insert(thread
->mg
, map
);
185 static int thread__clone_map_groups(struct thread
*thread
,
186 struct thread
*parent
)
190 /* This is new thread, we share map groups for process. */
191 if (thread
->pid_
== parent
->pid_
)
194 /* But this one is new process, copy maps. */
195 for (i
= 0; i
< MAP__NR_TYPES
; ++i
)
196 if (map_groups__clone(thread
->mg
, parent
->mg
, i
) < 0)
202 int thread__fork(struct thread
*thread
, struct thread
*parent
, u64 timestamp
)
206 if (parent
->comm_set
) {
207 const char *comm
= thread__comm_str(parent
);
210 err
= thread__set_comm(thread
, comm
, timestamp
);
215 thread
->ppid
= parent
->tid
;
216 return thread__clone_map_groups(thread
, parent
);
219 void thread__find_cpumode_addr_location(struct thread
*thread
,
220 enum map_type type
, u64 addr
,
221 struct addr_location
*al
)
224 const u8
const cpumodes
[] = {
225 PERF_RECORD_MISC_USER
,
226 PERF_RECORD_MISC_KERNEL
,
227 PERF_RECORD_MISC_GUEST_USER
,
228 PERF_RECORD_MISC_GUEST_KERNEL
231 for (i
= 0; i
< ARRAY_SIZE(cpumodes
); i
++) {
232 thread__find_addr_location(thread
, cpumodes
[i
], type
, addr
, al
);